Reuters: In Tesla Autopilot Probe, US Prosecutors Focus on Securities, Wire Fraud
Tesla’s Autopilot and Full Self-Driving systems assist with steering, braking and lane changes – but are not fully autonomous. While Tesla has warned drivers to stay ready to take over driving, the Justice Department is examining other statements by Tesla and Chief Executive Elon Musk suggesting its cars can drive themselves.
U.S. regulators have separately investigated hundreds of crashes, including fatal ones, that have occurred in Teslas with Autopilot engaged, resulting in a mass recall by the automaker.
Reuters exclusively reported the U.S. criminal investigation into Tesla in October 2022, and is now the first to report the specific criminal liability federal prosecutors are examining.
Investigators are exploring whether Tesla committed wire fraud, which involves deception in interstate communications, by misleading consumers about its driver-assistance systems, the sources said. They are also examining whether Tesla committed securities fraud by deceiving investors, two of the sources said.
The Securities and Exchange Commission is also investigating Tesla’s representations about driver-assistance systems to investors, one of the people said. The SEC declined to comment.
Tesla did not respond to a request for comment. Last October, it disclosed in a filing that the Justice Department had asked the company for information about Autopilot and Full Self-Driving.
The Justice Department declined to comment.
The probe, which is not evidence of wrongdoing, could result in criminal charges, civil sanctions, or no action. Prosecutors are far from deciding how to proceed, one of the sources said, in part because they are sifting through voluminous documents Tesla provided in response to subpoenas.
Reuters could not determine the specific statements prosecutors are reviewing as potentially illegal. Musk has aggressively touted the prowess of Tesla’s driver-assistance technology for nearly a decade.
Tesla videos demonstrating the technology that remain archived on its website say: “The person in the driver’s seat is only there for legal reasons. He is not doing anything. The car is driving itself.”
A Tesla engineer testified in 2022 in a lawsuit over a fatal crash involving Autopilot that one of the videos, posted in October 2016, intended to show the technology’s potential and did not accurately portray its capabilities at the time. Musk nevertheless posted the video on social media, writing: “Tesla drives itself (no human input at all) thru urban streets to highway streets, then finds a parking spot.”
In a conference call with reporters in 2016, Musk described Autopilot as “probably better” than a human driver. During an October 2022 call, Musk addressed a forthcoming FSD upgrade he said would allow customers to travel “to your work, your friend’s house, to the grocery store without you touching the wheel.”
Musk is increasingly focused on self-driving technology as Tesla’s car sales and profit slump. Tesla recently slashed costs through mass layoffs and shelved plans for a long-awaited $25,000 model that had been expected to drive sales growth.
“Going balls to the wall for autonomy is a blindingly obvious move,” the billionaire executive posted on his social-media platform X in mid-April. Tesla shares, down more than 28% so far this year, surged in late April when Musk visited China and made progress toward approvals to sell FSD there.
Musk has repeatedly promised self-driving Teslas for about a decade. “Mere failure to realize a long-term, aspirational goal is not fraud,” Tesla lawyers said in a 2022 court filing.
LEGAL CHALLENGES
Prosecutors scrutinizing Tesla’s autonomous-car claims are proceeding with caution, recognizing the legal hurdles they face, the people familiar with the inquiry said.
They will need to demonstrate that Tesla’s claims crossed a line from legal salesmanship to material and knowingly false statements that unlawfully harmed consumers or investors, three legal experts uninvolved in the probe told Reuters.
U.S. courts previously have ruled that “puffery” or “corporate optimism” regarding product claims do not amount to fraud. In 2008, a federal appeals court ruled that statements of corporate optimism alone do not demonstrate that a company official intentionally misled investors.
Justice Department officials will likely seek internal Tesla communications as evidence that Musk or others knew they were making false statements, said Daniel Richman, a Columbia Law School professor and former federal prosecutor. That is a challenge, Richman said, but the safety risk involved in overselling self-driving systems also “speaks to the seriousness with which prosecutors, a judge and jury would take the statements.”
FATAL CRASHES
Tesla’s claims about Autopilot and FSD have also drawn scrutiny in regulatory investigations and lawsuits.
Safety regulators and courts have raised concerns in recent months that corporate messaging about the technology – including the brand names Autopilot and Full Self-Driving – have imbued customers with a false sense of security.
In April, the Washington State Patrol arrested a man on suspicion of vehicular homicide after his Tesla, with Autopilot engaged, struck and killed a motorcyclist while the driver looked at his phone, police records show. In a probable-cause statement, a trooper cited the driver’s “admitted inattention to driving, while on autopilot mode … putting trust in the machine to drive for him.”
In Washington state, a driver remains “responsible for the safe and legal operation of that vehicle” regardless of its technological capabilities, a state patrol spokesperson told Reuters.
The same month, the U.S. National Highway Traffic Safety Administration launched an investigation into whether a Tesla recall of more than 2 million vehicles in December adequately addressed safety issues with Autopilot.
NHTSA declined to comment.
The recall followed a long-running probe opened by regulators after cars with Autopilot engaged repeatedly crashed into vehicles at first-responder emergency scenes. Regulators subsequently examined hundreds of crashes where Autopilot was engaged and identified 14 deaths and 54 injuries.
Tesla disputed NHTSA’s findings but agreed to the recall, which employed over-the-air software updates intended to alert inattentive drivers.
The NHTSA investigation found “a critical safety gap between drivers’ expectations” of Tesla’s technology “and the system’s true capabilities,” according to agency records. “This gap led to foreseeable misuse and avoidable crashes.”
(Reporting by Mike Spector and Chris Prentice in New YorkAdditional reporting by Hyunjoo Jin in San Francisco and David Shepardson in WashingtonEditing by Brian Thevenot and Matthew Lewis)