The safety driver behind the wheel of a self-driving Uber that struck and killed a woman in 2018 has been charged with a crime. Prosecutors in Maricopa County, Arizona, Tuesday said the driver, Rafaela Vasquez, has been indicted for criminal negligence. But Uber, her employer and the company that built the automated system involved in the fatal collision, won’t face charges.
The attorney for neighboring Yavapai County declined to prosecute Uber last year, writing in a letter that the office found “no basis for criminal liability.” (Yavapai took over the Uber part of the case because Maricopa County had worked with Uber on an anti-drunk-driving campaign.) Yavapai County attorney Sheila Polk declined to elaborate on her decision. A spokesperson for Uber declined to comment.
What happens when humans and machines work together to hurt others? The question isn’t new. As the anthropologist Madeleine Clare Elish noted earlier this year after an investigation into automation in the aviation sector, “conceptions of legal liability and responsibility did not adequately keep pace with advances in technology.” It has, in other words, been difficult—though not impossible—for the legal system to hold people responsible for the technology they build. Instead, the human in the loop, the person behind the wheel or the screen, has borne the bulk of the responsibility.
As a practical matter, it’s easier for prosecutors to sell juries on a story they already know. Vasquez was behind the wheel of a car and allegedly watching her cell phone instead of the darkened road in front of her when the car struck and killed a woman named Elaine Herzberg. People know about distracted driving. “That’s a simple story, that her negligence was the cause of [Herzberg’s] death,” says Ryan Calo, a law professor who studies robotics at the University of Washington School of Law. “Bring a case against the company, and you have to tell a more complicated story about how driverless cars work and what Uber did wrong.”
The story is more complicated, and more technical. Last year, the National Transportation Safety Board released its final report on the crash, the country’s first fatal one involving an autonomous vehicle. After combing through documents and software and interviews with Uber staffers, the safety panel determined that lots of people were responsible for the collision.
“Safety starts at the top,” NTSB chair Robert Sumwalt said. “The collision was the last link of a long chain of actions and decisions made by an organization that unfortunately did not make safety the top priority.” Among the culprits: Vasquez and Uber self-driving execs, who created what the NTSB called an “inadequate safety culture.”
social experiment by Livio Acerbo #greengroundit #wired https://www.wired.com/story/why-not-uber-charged-fatal-self-driving-car-crash