Credit: Inside Edition

We were promised a very near future where autonomous machines would be serving our needs and vehicle ownership would be rendered unnecessary: robots would quickly and efficiently deliver our orders and we could squeeze in a few more hours of work or sleep while being chauffeured around in self-driving cars.

Progress has been made, at least, on some of this. University campuses and cities across North America have indeed witnessed the growing presence of small food-delivery robots. Likewise, new partnerships have recently been announced to develop and test the safety of self-driving trucks.

The journey toward autonomous or self-driving consumer cars, on the other hand, has arguably come to a screeching halt. In 2021, top industry experts recognized that developing safe autonomous driving systems was not as simple as it was anticipated. Among them, Elon Musk himself conceded that developing the technology required to deliver safe self-driving cars has proved harder that he thought.

Automation paradox

More bad news came this week when the U.S. National Highway Traffic Safety Administration (NHTSA) released numbers that showed Tesla vehicles being responsible for nearly 70% of the crashes involving so-called SAE Level 2 cars.

Some cars are completely autonomous and are capable of driving without any input from the human driver. For example, Waymo One, in Phoenix, Ariz., is a ride-hailing service that currently deploys autonomous cars on a test route.

SAE Level 2 autonomous systems, like Tesla Autopilot, require human drivers to stay alert at all times, even when the system temporarily takes control of steering and acceleration. As soon as the traffic or aren't adequate for the system to operate, control is given back to the driver who needs to take over manual control of the .

Inside Edition looks at people’s behaviours in autonomous cars.

Human factors engineering is a cross-disciplinary research field investigating how humans interact with vehicle technology. Its researchers have, for years, highlighted the safety risks of automated driving—especially when the system requires the driver to make up for technological shortcomings to operate safely.

This is the case in what is known as the automation paradox, wherein the more automated the vehicle, the harder it is for humans to operate it properly.

Overestimating vehicle capability

Among the most prominent risks of operating SAE Level 2 cars is when drivers misunderstand the capabilities of the automated system. The issue often leads to unsafe behaviors like reading a book or taking a nap while the vehicle is in motion.

In 2021, there were so many reports of unsafe behaviors at the wheel of Level 2 cars, that the NHTSA required manufacturers to start reporting crashes that had occurred when these systems were engaged.

The initial findings, released in June 2022, showed that since 2021, Tesla and Honda vehicles were, respectively, involved in 273 and 90 reported crashes when these systems were engaged. Most crashes occurred in Texas and California.

While these data paint a dismal picture of the safety of these systems, they pale in comparison to the over 40,000 reported fatal crashes that occurred in the United States in 2021 alone.

As part of the same report, NHTSA itself highlights some of the methodological limitations of the study: from the incompleteness of some of the source data to failing to account for individual manufacturers' total vehicle volume or distance traveled by vehicles.

For the skeptics, this does not spell the end of . It does, however, confirm that widespread deployment of safe self-driving cars is not years, but decades, in the making.

Provided by The Conversation