Despite moving towards a technology reliant world, 95% of road traffic accidents are caused by human drivers either due to negligence on their parts or by some unknown circumstances.
According to the latest IDTechEx report “Autonomous Cars, Robotaxis & Sensors 2022–2042” it provides analysis on this data, spanning the last two and a half years’ worth of the reports. It is not looking good for human drivers.
Every company testing autonomous vehicles in California must report any collision to the California DMV.
From 187 reports, only 2 incidents could be attributed to the poor performance of the autonomous system. That means a staggering 99% of crashes involving autonomous vehicles are caused by human error. It is also worth mentioning that while the vehicles were operating in autonomous mode, 81 out of the 83 recorded incidents were caused by a human, either in another vehicle or as a misbehaving pedestrian.
The results, shown in the graphic, only tell half the story though. When delving into individual cases the ineptitude of some drivers becomes glaringly obvious. Here are four example cases where it would simply be impossible for two autonomous vehicles to have the same collision. These cases are summaries of real incidents documented and available on the California DMV website.
Case 1 (Pony.ai, 12th July 2019)
At a traffic light-controlled junction, the autonomous test driver notices the car in front has its reverse lights on. As a precaution, the test driver backs up 20-30 feet. When the lights turn green, the car in front accelerates in reverse, they manage to stop short of the Pony.ai vehicle by 4 feet. Without changing out of reverse, the driver in front accelerates again and hits the Pony.ai vehicle.
Case 2 (Cruise, 27th March 2021)
Here a Cruise vehicle was operating in autonomous mode, making a left turn at a 4-way junction when a vehicle behind attempted to overtake and carry straight on. The vehicle collided with the front left of the autonomous Cruise.
Case 3 (Cruise, 25th February 2021)
The Cruise vehicle, operating in autonomous mode was making a right turn at a junction and had right of way. A vehicle approaching from the left failed to stop for a stop sign and collided with the autonomous car.
Case 4 (Zoox, 22nd November 2019)
The autonomous vehicle was being driven by the human test driver in manual mode. The test driver entered a junction under a green light and was hit by another vehicle illegally entering the junction while fleeing from the local police force.
The cases mentioned here are some of the more unusual and have been shared to demonstrate how poor human driving can sometimes be. In the vast majority of the 81 cases where a human driver collided with an autonomously driven vehicle, the circumstances were far less interesting. The most common form of the collision was simply being rear-ended while in traffic or stopped. These kinds of crashes are likely caused by either human inattention or distraction.
The point remains though that in the cases given and most other cases, it is highly unlikely that a collision would have occurred if both were autonomous vehicles. Autonomous drivers don’t mistake forward and reverse gears, they don’t take unnecessary risks such as overtakes and running stop signs, they always pay 100% attention, and finally, they won’t flee from the police.
What about the two cases where the autonomous vehicle was at fault? Both cases involved a Zoox vehicle, and in both cases, the Zoox autonomous system appeared to misjudge its clearance to parked vehicles it was attempting to navigate around. The Zoox vehicles made contact causing some minor damage.
This was likely either a weakness or blind spot in the sensor suite, or a fault of the autonomous driver. Either way, as R&D activities are still underway, autonomous vehicles are not yet 100% infallible, and likely never will be. More information about the current maturity of robotaxi players can be found in IDTechEx’s latest report, “Autonomous Cars, Robotaxis & Sensors 2022–2042”.
Part of the problem could be assumed behavior, where the human driver expects the autonomous driver to go and they start moving in anticipation, only to hit the still stationary autonomous vehicle. Like being caught out by a learner driver not moving at a roundabout or other junction; sometimes there looks to be a sufficient gap and the driver begins to move, expecting that the learner has gone for it, but the overly cautious learner is still there.
To further cement this point, while studying disengagement reports from robotaxi leader Waymo, IDTechEx found that approximately 20% of disengagements were caused by the test driver reacting to a nearby human driver’s poor behavior.
This could be failing to abide by the rules and etiquette of the road, or other drivers behaving aggressively towards the autonomous vehicle. Similar incidents were recorded in the collision reports, for example, a couple of pedestrians attacked two Cruise vehicles with wooden sticks in a completely unprovoked display of aggression.
Autonomous and connected vehicles have many advantages over human drivers. They have permanent 360° perception, they can communicate their intentions to each other in advance, and they do not make silly operational errors (such as accidentally engaging reverse). As their maturity progresses it will be hard to deny that they are superior to human drivers, and we will need to take a back seat in the task of driving.
For more details, please click here.