People fought against seatbelts not that long ago. Though, it was long enough that an “anti-seatbelt movement” sounds absurd to us at this point in time. Cruise control got a lot of backlash a couple of decades ago as well, but now it’s in basically every new car and no one bats an eye. There was a big anti-cellphone movement when cellphones started getting popular. Who doesn’t have a cellphone today?
It doesn’t matter what has happened in the past, many humans are naturally scared of new things, and that includes new technologies. That ain’t gonna change — it’s embedded deep in our DNA. But that doesn’t mean that we all have to accept illogical fear and that all new technologies will be stifled by fearful humans stomping on them. Clearly, technology keeps evolving and rolling out. However, some tech does face counterproductive obstructionism that can really slow progress, even progress that can help save lives.
On the opposite side of the coin from our fear of new things, if something has been happening for a long time, or even obviously plaguing us for a long time, humans have a separate ability/tendency to just basically accept it. It’s normal. Even if it’s deadly. A recent tweet from one of my favorite tweeters, Not_an_Analyst, pointed this out exceptionally well.
Human pilot had a terrible 2020, with death rates on the roads up 23.4% per vehicle mile traveled
7340 excess deaths compared to 2019’s fatality rate per VMT
But Autopilot is the problem, if you believe the press and politicians pic.twitter.com/juFjs2cqoE
— Not_an_Analyst (@facts_tesla) August 19, 2021
The clever wording here is of course referencing Tesla Autopilot, and the alternative to Autopilot — “Humanpilot.” If one person dies a year in a crash involving Autopilot, many in the media go crazy, which triggers backlash well beyond the media. Yes, the loss of any life is tragic, but that doesn’t make the calculation 1–0. In 2020, 23,395 passenger vehicle occupants died in the US. That’s more than 64 people a day dying. Essentially all of those deaths could be called “Humanpilot deaths” — if we really want to put Tesla Autopilot into comparative context.
Perhaps there was one or maybe two Autopilot-related deaths in the past year, but I can’t find any such stories (I just went through our Autopilot archives for the past year to make sure my memory wasn’t betraying me). Also, I’m confident that we’d hear about an Autopilot-related death if there was one. If, indeed, there were no vehicle passenger deaths in the past year from a Tesla with Autopilot engaged, that would mean the actual “score” was: Autopilot 0 — Humanpilot 23,395.
If we all had Autopilot on our cars by default and some company was introducing or expanding the use of “Humanpilot,” is there any chance that effort would succeed? Yes, these are absolute numbers, not rates of death, but if that zero was the actual total last year, then Autopilot’s rate of death was 0%, whereas Humanpilot’s rate was clearly much higher.
Wreck technician James Law told us this week that there have been 38 deaths this year in the US from cars striking people or cars at emergency vehicle scenes, and none of those have been deaths caused by Tesla vehicles.
There’s approximately one crash every 4 million miles for a Tesla driver with Autopilot engaged. The national average is approximately one crash every 484,000 miles. That means that, relatively speaking, for every 1 Autopilot-related crash, there are 8.7 auto crashes across the US. Yes, there are flaws with this data comparison — the classes of Tesla vehicles are above average, and higher class vehicles may get into crashes less in general; the Tesla fleet is much younger than the overall auto fleet; more Autopilot miles are driven on highways, relatively speaking, and there’s reportedly (and logically) a lower rate of crash per mile on highways than in cities. Nonetheless, 1 to 8.7 is a tremendous ratio.
The irony here is that there are certain people calling for Tesla Autopilot to be banned, investigated, and sent straight to Hell — do not pass go, do not collect $100.
There are an estimated 328,000 crashes a year in the US caused by sleepy drivers. People can’t just go out and driver with Tesla Autopilot while falling asleep or drunk and expect to make it to their destination, but recent stories have shown that Autopilot can indeed get such a person stopped and rescued safely before something tragic happens. Here’s a case from Norway, for example. It’s hard to deny that Autopilot helped in this case. It’s easy to imagine that driver dying or causing another person’s death if he was in another car. There’s also this case of a Wisconsin driver who fell asleep while driving and also didn’t even end up crashing, thanks to Autopilot. That’s not to say every such story will end well, but “Humanpilot” apparently leads to 328,000 crashes a year from drowsy drivers, and I have to wonder how many of those crashes would have been averted if Autopilot was on when the drivers were dozing off.
Two people died in a Tesla in the Houston area earlier this year. There were quick claims that Autopilot was on, leading to widespread calls (from both the political right and the political left) to investigate Tesla over this and perhaps shut down Autopilot in the meantime. It turns out that Autopilot wasn’t even on. “Nonetheless, that’s irrelevant — investigate Autopilot” seems to be the call to action anyway. There wasn’t a connection between the two in reality, but impression that there was — in this one case — is enough to stimulate or increase the pressure to investigate Tesla Autopilot.
It’s hard to see how anyone neutral and objective here could look at this situation and not think it’s a bit absurd. Everything I’ve seen indicates that Autopilot helps to save lives, and we all know that “Humanpilot” is an absolute disaster with a horrible safety record, yet people who supposedly want to protect lives are up in arms about the need to investigate and shut down Autopilot. As I wrote in April, “Tesla Autopilot Is Just Better Cruise Control — Anyone Who Thinks It Should Be Banned Is Acting Stupid.” Fake controversy and manufactured fear from a non-Autopilot accident is not a good enough reason to launch an all-out offensive on an advanced driver assistance system (ADAS) that is saving lives. Let’s not go down that road.
Appreciate CleanTechnica’s originality? Consider becoming a CleanTechnica Member, Supporter, Technician, or Ambassador — or a patron on Patreon.