NTSB: Autopilot Was in Use Before Tesla Hit Semitrailer

Tesla
A Tesla Model 3. (Patrick T. Fallon/Bloomberg News)

DETROIT — A Tesla Model 3 involved in a fatal crash with a semitrailer in Florida on March 1 was operating on the company’s semi-autonomous Autopilot system, federal investigators have determined.

The car drove beneath the trailer, killing the driver, in a crash that is strikingly similar to one that happened on the other side of Florida in 2016 that also involved use of Autopilot.

In both cases, neither the driver nor the Autopilot system stopped for the trailers, and the roofs of the cars were sheared off.

The crash, which remains under investigation by the National Transportation Safety Board and the National Highway Traffic Safety Administration, raises questions about the effectiveness of Autopilot, which uses cameras, long-range radar and computers to detect objects in front of the cars to avoid collisions. The system also can keep a car in its lane, change lanes and navigate freeway interchanges.



Tesla has maintained that the system is designed only to assist drivers, who must pay attention at all times and be ready to intervene.

Image

In a preliminary report on the March 1 crash, NTSB said that data and video from the Tesla show that the driver turned on Autopilot about 10 seconds before the crash on a divided highway with turn lanes in the median. From less than eight seconds until the time of the crash, the driver’s hands were not detected on the steering wheel, NTSB report stated.

“Neither the preliminary data nor the videos indicate that the driver or the ADAS [Advanced Driver Assist System] executed evasive maneuvers,” the report stated.

The Model 3 was going 68 mph when it hit the trailer on U.S. 441, the report said. Jeremy Beren Banner, 50, was killed.

Tesla said in a statement May 16 that Banner did not use Autopilot at any other time during the drive before the crash. Vehicle logs show that he took his hands off the steering wheel immediately after activating Autopilot, the statement said.

Tesla also said it is saddened by the crash and that drivers have traveled more than 1 billion miles while using Autopilot.

“When used properly by an attentive driver who is prepared to take control at all times, drivers supported by Autopilot are safer than those operating without assistance,” the company said.

The circumstances of the Delray Beach crash are much like one that occurred in May 2016 near Gainesville, Florida. Joshua Brown, 40, of Canton, Ohio, was traveling in a Tesla Model S on a divided highway and using the Autopilot system when he was killed.

Image

Musk

Neither Brown nor the car braked for a tractor-trailer, which had turned left in front of the Tesla and was crossing its path. Brown’s Tesla also went beneath the trailer and its roof was sheared off. After that crash, Tesla CEO Elon Musk said the company made changes in its system so radar would play more of a role in detecting objects.

David Friedman, who was acting head of NHTSA in 2014 and now is vice president of advocacy for Consumer Reports, said he was surprised the agency didn’t declare Autopilot defective after the Gainesville crash and seek a recall. The Delray Beach crash, he said, reinforces that Autopilot is being allowed to operate in situations that it cannot handle safely. “Their system cannot literally see the broad side of an 18-wheeler on the highway,” Friedman said.

Tesla’s system was too slow to warn the driver to pay attention, unlike systems that Consumer Reports has tested from General Motors Co. and other companies, Friedman said. GM’s Super Cruise driver assist system only operates on divided highways with no median turn lanes, he said.

Tesla needs a better system to more quickly detect whether drivers are paying attention and warn them if they are not, Friedman said. “Tesla has for too long been using human drivers as guinea pigs. This is tragically what happens,” he said.

To force a recall, NHTSA must do an investigation and show that the way a vehicle is designed is outside of industry standards. “There are multiple systems out on the roads right now that take over some level of steering and speed control, but there’s only one of them that we keep hearing about where people are dying or getting into crashes. That kind of stands out,” he said.