The federal government’s top auto-safety agency is significantly expanding an investigation into Tesla and its Autopilot driver-assistance system to determine if the technology poses a safety risk.
The agency, the National Highway Traffic Safety Administration, said Thursday that it was upgrading its preliminary evaluation of Autopilot to an engineering analysis, a more intensive level of scrutiny that is required before a recall can be ordered.
The analysis will look at whether Autopilot fails to prevent drivers from diverting their attention from the road and engaging in other predictable and risky behavior while using the system.
“We’ve been asking for closer scrutiny of Autopilot for some time,” said Jonathan Adkins, executive director of the Governors Highway Safety Association, which coordinates state efforts to promote safe driving.
NHTSA said it had not determined whether Autopilot has defects that can cause cars to crash while it is engaged.
The wider investigation covers 830,000 vehicles sold in the United States. They include all four Tesla cars – the Models S, X, 3 and Y – in model years from 2014 to 2021.
The agency’s preliminary evaluation focused on 11 incidents in which Tesla cars operating under Autopilot control crashed into emergency vehicles that were parked and had their lights flashing. In that review, NHTSA said Thursday, the agency became aware of 191 crashes – not limited to those involving emergency vehicles – that warranted closer investigation.
Tesla did not immediately respond to a request for comment on the agency’s move.
The deepening of the investigation signals that NHTSA is more seriously considering safety concerns stemming from a lack of safeguards to prevent drivers from using Autopilot in a dangerous manner.
“This isn’t your typical defect case,” said Michael Brooks, acting executive director at the Center for Auto Safety, a nonprofit consumer advocacy group. “They are actively looking for a problem that can be fixed, and they’re looking at driver behavior, and the problem may not be a component in the vehicle.”
Competing systems developed by General Motors and Ford Motor use infrared cameras that closely track the driver’s eyes and sound warning chimes if a driver looks away from the road for more than two or three seconds. Tesla did not initially include such a driver monitoring system in its cars, and only later added a standard camera that is much less precise than infrared cameras in eye tracking.
Tesla tells drivers to use Autopilot only on divided highways, but the system can be activated on any streets that have lines down the middle. The GM and Ford systems – known as Super Cruise and BlueCruise – can be activated only on highways.
Autopilot was first offered in Tesla models in late 2015. It uses cameras and other sensors to steer, accelerate and brake with little input from drivers. Owner manuals tell drivers to keep their hands on the steering wheel and their eyes on the road, but early versions of the system allowed drivers to keep their hands off the wheel for five minutes or more under certain conditions.
Questions about the system arose in 2016 when an Ohio man was killed when his Tesla Model S crashed into a tractor-trailer on a highway in Florida while Autopilot was activated. NHTSA investigated that crash and in 2017 said it had found no safety defect in Autopilot.
The Issues With Tesla’s Autopilot System
Claims of safer driving. Tesla cars can use computers to handle some aspects of driving, such as changing lanes. But there are concerns that this driver-assistance system, called Autopilot, is not safe. Here is a closer look at the issue.
Since then, more Autopilot crashes and fatalities have occurred. NHTSA has said it is aware of 35 crashes that occurred while Autopilot was activated, including nine that resulted in the deaths of 14 people.
NHTSA opened its preliminary evaluation of Autopilot last August and initially focused on 11 incidents in which Teslas operating with Autopilot engaged engaged into police cars, fire trucks and other emergency vehicles that had stopped and had their lights flashing. Those crashes resulted in one death and 17 injuries.
While examining those incidents, it discovered six more crashes with emergency vehicles and eliminated one of the original 11 from further study.
At the same time, the agency also learned of dozens of more crashes that occurred while Autopilot was active and did not involve emergency vehicles. Of those, the agency first focused on 191, and eliminated 85 from further scrutiny because it could not obtain enough information to get a clear picture if Autopilot was a major cause.
In about half of the remaining 106, NHTSA found evidence that suggested drivers did not have their full attention on the road. About a quarter of the 106 occurred on roads where Autopilot is not supposed to be used.
In an engineering analysis, NHTSA typically acquires vehicles it is examining and puts them through rigorous testing to try to identify flaws and replicate problems they can cause. In the past it has taken apart components to find faults, and has asked manufacturers for detailed data on how components operate, often including proprietary information.
It can take NHTSA months or even a year or more to complete the process.