The sun had yet to rise in Delray Beach, Fla., when Jeremy Banner flicked on Autopilot. His red Tesla Model 3 sped down the highway at nearly 70 mph, his hands no longer detected on the wheel.
Seconds later, the Tesla plowed into a semi-truck, shearing off its roof as it slid under the truck’s trailer. Banner was killed on impact.
Banner’s family sued after the gruesome 2019 collision, one of at least 10 active lawsuits involving Tesla’s Autopilot, several of which are expected to go to court over the next year. Together, the cases could determine whether the driver is solely responsible when things go wrong in a vehicle guided by Autopilot — or whether the software should also bear some of the blame.
The outcome could prove critical for Tesla, which has pushed increasingly capable driver-assistance technology onto the nation’s roadways far more rapidly than any other major carmaker. If Tesla prevails, the company could continue deploying the evolving technology with few legal consequences or regulatory guardrails. Multiple verdicts against the company, however, could threaten both Tesla’s reputation and its financial viability.
According to an investigation by the National Transportation Safety Board (NTSB), Banner, a 50-year-old father of four, should have been watching the road on that March morning. He agreed to Tesla’s terms and conditions of operating on Autopilot and was provided with an owner’s manual, which together warn of the technology’s limitations and state that the driver is ultimately responsible for the trajectory of the car.
But lawyers for Banner’s family say Tesla should shoulder some responsibility for the crash. Along with former transportation officials and other experts, they say the company’s marketing of Autopilot exaggerates its capabilities, creating a false sense of complacency that can lead to deadly crashes. That argument is echoed in several Autopilot-related cases, where plaintiffs say they believed Tesla’s claims that Autopilot was “safer than a human-operated vehicle.”
A Washington Post analysis of federal data found that vehicles guided by Autopilot have been involved in more than 700 crashes, at least 19 of them fatal, since its introduction in 2014, including the Banner crash. In Banner’s case, the technology failed repeatedly, his family’s lawyers argue, from when it didn’t brake to when it didn’t issue a warning about the semi-truck in the car’s path.
To reconstruct the crash, The Post relied on hundreds of court documents, dash cam photos and a video of the crash taken from a nearby farm, as well as satellite imagery, NTSB crash assessment documents and diagrams, and Tesla’s internal data log, which the NTSB included in its investigation report. The Post’s reconstruction found that braking just 1.6 seconds before impact could have avoided the collision.
Friday, March 1, 2019, starts like any workday for Banner, a software engineer who heads to work in his 2018 Tesla Model 3 around 5:50 a.m.
At 6:16 a.m., Banner sets cruise control to a maximum of 69 mph, though the speed limit on U.S. 441 is 55. He turns on Autopilot 2.4 seconds later.
A standard Autopilot notice flashes on the screen: “Please keep your hands on the wheel. Be prepared to take over at any time.”
According to Tesla’s user documentation, Autopilot wasn’t designed to work on a highway with cross-traffic such as U.S. 441. But drivers sometimes can activate it in areas and under conditions for which it is not designed.
Two seconds later, the Tesla’s data log registers no “driver-applied wheel torque,” meaning Banner’s hands cannot be detected on the wheel.
If Autopilot does not detect a driver’s hands, it flashes a warning. In this case, given Banner’s speed, the warning would have come after about 25 seconds, according to the NTSB investigation.
Banner does not have that long.
From a side road, a truck driver begins to cross U.S. 441, slowing but failing to fully stop at a stop sign.
The truck enters the Tesla’s lane of traffic.
Two seconds later — just before impact — the Tesla’s forward-facing camera captures this image of the truck.
The car does not warn Banner of the obstacle. “According to Tesla, the Autopilot vision system did not consistently detect and track the truck as an object or threat as it crossed the path of the car,” the NTSB crash report says.
The Tesla continues barreling toward the tractor-trailer at nearly 69 mph. Neither Banner nor Autopilot activates the brakes.
The Tesla slams into the truck, and its roof is ripped off as it passes under the trailer. Banner is killed instantly.
The Tesla continues on for another 40 seconds, traveling about 1,680 feet — nearly a third of a mile — before finally coasting to a stop on a grassy median.
A surveillance video located on the farm where the truck driver had just made a routine delivery shows the crash in real time. This video, which was obtained exclusively by The Post, along with court documents, crash reports and witness statements, offers a rare look at the moments leading up to an Autopilot crash. Tesla typically does not provide access to its cars’ crash data and often prevents regulators from revealing crash information to the public.
Braking even 1.6 seconds before the crash could have avoided the collision, The Post’s reconstruction found by reviewing braking distance measurements of a 2019 Tesla Model 3 with similar specifications, conducted by vehicle testers at Car and Driver. At this point the truck was well within view and spanning both lanes of southbound traffic.
Tesla braking distance map
Due to the uncertainty of Banner’s movements in the car, The Post did not depict him in the reconstruction. The NTSB investigation determined that Banner’s inattention and the truck driver’s failure to fully yield to oncoming traffic were probable causes of the crash.
However, the NTSB also cited Banner’s “overreliance on automation,” saying Telsa’s design “permitted disengagement by the driver” and contributed to the crash. Four years later, despite pleas from safety investigators, regulators in Washington have outlined no clear plan to address those shortcomings, allowing the Autopilot experiment to continue to play out on American roads, with little federal intervention.
While the Federal Motor Vehicle Safety Standards administered by the National Highway Traffic Safety Administration (NHTSA) spell out everything from how a car’s brakes should operate to where its lights should be located, they offer little guidance about vehicle software.
‘Fancy cruise control’
Teslas guided by Autopilot have slammed on the brakes at high speeds without clear cause, accelerated or lurched from the road without warning and crashed into parked emergency vehicles displaying flashing lights, according to investigation and police reports obtained by The Post.
In February, a Tesla on Autopilot smashed into a firetruck in Walnut Creek, Calif., killing the driver. The Tesla driver was under the influence of alcohol during the crash, according to the police report.
In July, a Tesla rammed into a Subaru Impreza in South Lake Tahoe, Calif. “It was, like, head on,” according to a 911 call from the incident obtained by The Post. “Someone is definitely hurt.” The Subaru driver later died of his injuries, as did a baby in the back seat of the Tesla, according to the California Highway Patrol.
Tesla did not respond to multiple requests for comment. In its response to the Banner family’s complaint, Tesla said, “The record does not reveal anything that went awry with Mr. Banner’s vehicle, except that it, like all other automotive vehicles, was susceptible to crashing into another vehicle when that other vehicle suddenly drives directly across its path.”
Autopilot includes features to automatically control the car’s speed, following distance, steering and some other driving actions, such as taking exits off a freeway. But a user manual for the 2018 Tesla Model 3 reviewed by The Post is peppered with warnings about the software’s limitations, urging drivers to always pay attention, with hands on the wheel and eyes on the road. Before turning on Autosteer — an Autopilot feature — for the first time, drivers must click to agree to the terms.
In particular, Tesla noted in court documents for the Banner case that Autopilot was not designed to reliably detect cross-traffic, or traffic moving perpendicular to a vehicle, arguing that its user terms offers adequate warning of its limitations.
In a Riverside, Calif., courtroom last month in a lawsuit involving another fatal crash where Autopilot was allegedly involved, a Tesla attorney held a mock steering wheel before the jury and emphasized that the driver must always be in control.
Autopilot “is basically just fancy cruise control,” he said.
Tesla CEO Elon Musk has painted a different reality, arguing that his technology is making the roads safer: “It’s probably better than a person right now,” Musk said of Autopilot during a 2016 conference call with reporters.
Musk made a similar assertion about a more sophisticated form of Autopilot called Full Self-Driving on an earnings call in July. “Now, I know I’m the boy who cried FSD,” he said. “But man, I think we’ll be better than human by the end of this year.”
The NTSB said it has repeatedly issued recommendations aiming to prevent crashes associated with systems such as Autopilot. “NTSB’s investigations support the need for federal oversight of system safeguards, foreseeable misuse, and driver monitoring associated with partial automated driving systems,” NTSB spokesperson Sarah Sulick said in a statement.
NHTSA said it has an “active investigation” of Autopilot. “NHTSA generally does not comment on matters related to open investigations,” NHTSA spokeswoman Veronica Morales said in a statement. In 2021, the agency adopted a rule requiring carmakers such as Tesla to report crashes involving their driver-assistance systems.
Beyond the data collection, though, there are few clear legal limitations on how this type of advanced driver-assistance technology should operate and what capabilities it should have.
“Tesla has decided to take these much greater risks with the technology because they have this sense that it’s like, ‘Well, you can figure it out. You can determine for yourself what’s safe’ — without recognizing that other road users don’t have that same choice,” former NHTSA administrator Steven Cliff said in an interview.
“If you’re a pedestrian, [if] you’re another vehicle on the road,” he added, “do you know that you’re unwittingly an object of an experiment that’s happening?”
‘The car is driving itself’
Banner researched Tesla for years before buying a Model 3 in 2018, his wife, Kim, told federal investigators. Around the time of his purchase, Tesla’s website featured a video showing a Tesla navigating the curvy roads and intersections of California while a driver sits in the front seat, hands hovering beneath the wheel.
The video, recorded in 2016, is still on the site today.
“The person in the driver’s seat is only there for legal reasons,” the video says. “He is not doing anything. The car is driving itself.”
In a different case involving another fatal Autopilot crash, a Tesla engineer testified that a team specifically mapped the route the car would take in the video. At one point during testing for the video, a test car crashed into a fence, according to Reuters. The engineer said in a deposition that the video was meant to show what the technology could eventually be capable of — not what cars on the road could do at the time.
While the video concerned Full Self-Driving, which operates on surface streets, the plaintiffs in the Banner case argue Tesla’s “marketing does not always distinguish between these systems.”
Not only is the marketing misleading, plaintiffs in several cases argue, the company gives drivers a long leash when deciding when and how to use the technology. Though Autopilot is supposed to be enabled in limited situations, it sometimes works on roads it’s not designed for. It also allows drivers to go short periods without touching the wheel and to set cruising speeds well above posted speed limits.
For example, Autopilot was not designed to operate on roads with cross-traffic, Tesla lawyers say in court documents for the Banner case. The system struggles to identify obstacles in its path, especially at high speeds. The stretch of U.S. 441 where Banner crashed was “clearly outside” the environment Autopilot was designed for, the NTSB said in its report. Still, Banner was able to activate it.
Identifying semi-trucks is a particular deficiency that engineers have struggled to solve since Banner’s death, according to a former Autopilot employee who spoke on the condition of anonymity for fear of retribution.
Tesla tasked image “labelers” with repeatedly identifying images of semi-trucks perpendicular to Teslas to better train its software “because even in 2021 that was a heavy problem they were trying to solve,” the former employee said.
Because of the orientation of Tesla’s cameras, the person said, it was sometimes hard to discern the location of the tractor-trailers. In one view, the truck could appear to be floating 20 feet above the road, like an overpass. In another view, it could appear 25 feet below the ground.
Tesla complicated the matter in 2021 when it eliminated radar sensors from its cars, The Post previously reported, making vehicles such as semi-trucks appear two-dimensional and harder to parse.
In 2021, the chair of the NTSB publicly criticized Tesla for allowing drivers to turn on Autopilot in inappropriate locations and conditions — citing Banner’s crash and a similar wreck that killed another man, Joshua Brown, in 2016.
A third similar crash occurred this past July, killing a 57-year-old bakery owner in Fauquier County, Va., after his Tesla collided with a semi-truck.
Philip Koopman, an associate professor at Carnegie Mellon who has studied self-driving-car safety for more than 25 years, said the onus is on the driver to understand the limitations of the technology. But, he said, drivers can get lulled into thinking the technology works better than it does.
“If a system turns on, then at least some users will conclude it must be intended to work there,” Koopman said. “Because they think if it wasn’t intended to work there, it wouldn’t turn on.”
Andrew Maynard, a professor of advanced technology transitions at Arizona State University, said customers probably just trust the technology.
“Most people just don’t have the time or ability to fully understand the intricacies of it, so at the end they trust the company to protect them,” he said.
It is impossible to know what Banner was doing in the final seconds of his life, after his hands were no longer detected on the wheel. Tesla has argued in court documents that if he had been paying attention to the road, it is “undisputed” that “he could have avoided the crash.”
The case, originally set for trial this week in Palm Beach County Circuit Court, has been delayed while the court considers the family’s request to seek punitive damages against Tesla.
A small jolt
Whatever the verdict, the crash that March morning had a shattering effect on the truck driver crossing U.S. 441. The 45-year-old driver — whom The Post is not naming because he was not charged — felt a small jolt against the back of his truck as Banner’s Tesla made impact. He pulled over and hopped out to see what had happened.
According to a transcript of his interview with the NTSB, it was still dark and difficult to see when the crash occurred. But the driver noticed pink-stained glass stuck on the side of his trailer.
“Are you the guy that drives this tractor?” he recalled a man in a pickup hollering.
“Yeah,” the driver said he responded.
“That dude didn’t make it,” the man told him.
The truck driver started to shake.
He said he should have been more careful at the stop sign that morning, according to an interview with federal investigators. Banner’s family also sued the driver, but they settled, according to the Banner family’s lawyer.
The truck driver told investigators that self-driving vehicles have always made him uneasy and that he doesn’t think they should be allowed on the road. He became emotional recounting the crash.
“I’ve done it a dozen times,” the driver said of his fateful left turn. “And I clearly thought I had plenty of time. I mean, it was dark, and the cars looked like they was back further than what they was.”
“Yeah,” the investigator said.
“And, I mean, it’s just something I’m —,” the driver said.
“It’s okay, it’s okay,” the investigator responded.
“Yeah, take your time,” another investigator said.
“Just,” the driver said, pausing again. “It’s something I’m going to have to live with.”
Methodology
To reconstruct Banner’s crash, The Post relied on hundreds of court documents, dash cam photos and a video of the crash taken from a nearby farm, as well as satellite imagery, NTSB assessment documents and diagrams, and Tesla’s internal data log. Speeds included in the Tesla’s data log were used by The Post to plot and animate the movement of the Tesla vehicle within a 3D model of the highway produced from OpenStreetMap data and satellite imagery. The Post used other visual material, such as diagrams, dash cam stills and a surveillance video of the crash, to further clarify the changing positions of the Tesla and plot the movement of the truck. The Tesla’s data log also included information on when certain system and Autopilot features were activated or not activated, which The Post time-coded and added into the animation to present the sequence of system events before the crash.
The Tesla interface featured in the animation is based upon the default display in a Tesla Model 3.
About this story
Additional research by Alice Crites and Monika Mathur. Editing by Christina Passariello, Karly Domb Sadof, Laura Stevens, Nadine Ajaka and Lori Montgomery. Copy-editing by Carey L. Biron.