Skip to main contentSkip to navigationSkip to navigation
A red Tesla Model X with the logo.
The driver, Kevin George Aziz Riad, 27, has pleaded not guilty. Riad, a limousine service driver, is free on bail while the case is pending. Photograph: David Zalubowski/AP
The driver, Kevin George Aziz Riad, 27, has pleaded not guilty. Riad, a limousine service driver, is free on bail while the case is pending. Photograph: David Zalubowski/AP

Tesla driver charged with vehicular manslaughter over fatal Autopilot crash

This article is more than 2 years old

Misuse of Autopilot, which can control steering, speed and braking, is the subject of investigations by two federal agencies

California prosecutors have filed two counts of vehicular manslaughter against the driver of a Tesla on Autopilot who ran a red light, slammed into another car and killed two people in 2019.

The defendant appears to be the first person to be charged with a felony in the United States for a fatal crash involving a motorist who was using a partially automated driving system. Los Angeles county prosecutors filed the charges in October, but they came to light only last week.

The driver, Kevin George Aziz Riad, 27, has pleaded not guilty. Riad, a limousine service driver, is free on bail while the case is pending.

The misuse of Autopilot, which can control steering, speed and braking, has occurred on numerous occasions and is the subject of investigations by two federal agencies.

The filing of charges in the California crash could serve notice to drivers who use systems like Autopilot that they cannot rely on them to control vehicles.

The criminal charges aren’t the first involving an automated driving system, but they are the first to involve a widely used driver technology.

Authorities in Arizona filed a charge of negligent homicide in 2020 against a driver Uber had hired to take part in the testing of a fully autonomous vehicle on public roads. The Uber vehicle, an SUV with the human backup driver on board, struck and killed a pedestrian.

By contrast, Autopilot and other driver-assist systems are widely used on roads across the world. An estimated 765,000 Tesla vehicles are equipped with it in the United States alone.

In the Tesla crash, police said a Model S was moving at a high speed when it left a freeway and ran a red light in the Los Angeles suburb of Gardena and struck a Honda Civic at an intersection on 29 December 2019. Two people who were in the Civic, Gilberto Alcazar Lopez and Maria Guadalupe Nieves-Lopez died at the scene. Riad and a woman in the Tesla were hospitalized with non-life threatening injuries.

Criminal charging documents do not mention Autopilot. But the National Highway Traffic Safety Administration, which sent investigators to the crash, confirmed last week that autopilot was in use in the Tesla at the time of the crash.

Riad’s defense attorney did not respond to requests for comment last week, and the Los Angeles county district attorney’s office declined to discuss the case.

Tesla did not immediately respond to a request for comment.

Tesla has updated its Autopilot software to try to make it harder for drivers to abuse it. It’s also tried to improve autopilot’s ability to detect emergency vehicles.

The company has said that Autopilot and a more sophisticated Full Self-Driving system cannot drive themselves and that drivers must pay attention and be ready to react at anytime. Full Self-Driving is being tested by hundreds of Tesla owners on public roads in the US.

Tesla did not immediately respond to a request for comment.

Bryant Walker Smith, a law professor at the University of South Carolina who studies automated vehicles, said this is the first US case to his knowledge in which serious criminal charges were filed in a fatal crash involving a partially automated driver-assist system. Tesla, he said, could be “criminally, civilly or morally culpable” if it is found to have put a dangerous technology on the road.

Donald Slavik, a Colorado lawyer who has served as a consultant in automotive technology lawsuits, including many against Tesla, said he, too, is unaware of any previous felony charges being filed against a US driver who was using partially automated driver technology involved in a fatal crash.

The families of Lopez and Nieves-Lopez have sued Tesla and Riad in separate lawsuits. They have alleged negligence by Riad and have accused Tesla of selling defective vehicles that can accelerate suddenly and that lack an effective automatic emergency braking system. A joint trial is scheduled for mid-2023.

Lopez’s family, in court documents, alleges that the car “suddenly and unintentionally accelerated to an excessive, unsafe and uncontrollable speed”.

Nieves-Lopez’s family further asserts that Riad was an unsafe driver, with multiple moving infractions on his record, and couldn’t handle the high-performance Tesla.


Asked about the manslaughter charges against Riad, the NHTSA issued a statement saying there is no vehicle on sale that can drive itself. And whether or not a car is using a partially automated system, the agency said, “every vehicle requires the human driver to be in control at all times”.

NHTSA added that all state laws hold human drivers responsible for operation of their vehicles. Though automated systems can help drivers avoid crashes, the agency said, the technology must be used responsibly.

Rafaela Vasquez, the driver in the Uber autonomous test vehicle, was charged in 2020 with negligent homicide after the SUV fatally struck a pedestrian in suburban Phoenix in 2018. Vasquez has pleaded not guilty. Arizona prosecutors declined to file criminal charges against Uber.

Most viewed

Most viewed