Autonomous Cars: Will They Become Mainstream in 2024?

Admin October 17, 2024

Autonomous Cars: Will They Become Mainstream in 2024?

The automotive industry is going through the most transformative journey in its history. Only 15 years ago, autonomous cars were a pipe dream; electric cars were the butt of jokes; and safety was reactive, aiming to minimize the consequences of collisions rather than prevent them. Now, in 2024, autonomous mobility services are available to the public across the U.S. and China; electric cars are a desirable mainstream option; and the future of safety is preventative, trying to ensure crashes never happen. This article considers three critical aspects of future vehicles: sensors, software and safety.

Sensors

Radar, cameras and LiDAR form the core trio of sensors synonymous with autonomous driving. All three have been evolving in the automotive industry and continue to progress. LiDAR is coming down to a price point more palatable for OEMs, radar’s performance and utility continues to evolve, and high-resolution camera technologies pioneered for smartphone cameras are being leveraged in autonomous driving systems.

Radar is perhaps the most interesting of the three categories. Automotive radar has come a long way since it was introduced to vehicles around the turn of the century. Back then, radar was good for detecting the distance to the vehicle ahead, and that was about it. Its major advantage was that it worked under almost any weather conditions the car might encounter. Snow, rain, fog, sandstorm—the radar did not care. For fully autonomous driving, however, its resolution simply was not sufficient.

In recent years, advances in semiconductor technology and the emerging demands of the autonomous driving industry have forced radar to evolve. It has sprouted more transmitter and receiver channels (Tx × Rx), growing from 1 × 2 to 3 × 4, and then to 12 × 16 from leading Tier 1 companies and 48 × 48 arrays from leading startups. Large antenna arrays with many transmitting and receiving channels have boosted the resolution of modern radar.

However, transmitting and receiving channels are like pixels on a camera: Simply adding more pixels does not necessarily make a better camera. As far back as 2013, the Nokia Lumina 1020 had 41-MP cameras, yet the iPhone 14 from 2022 uses 12-MP sensors. Which would take a better photo? There is more at play than just the pixel count. Sensitivity, dynamic range and post-processing software make a significant difference to the end image. Likewise, radar performance is not just a measure of transmitters and receivers.

An emerging priority for radar is dynamic range, or the ability of radar to detect both high-reflectivity objects and low-reflectivity objects simultaneously. Consider, for example, a small child standing near a car. The reflection from the car is orders of magnitude brighter than that from the child. For the radar, detecting the child is like trying photograph a grain of rice in front of the sun. It’s very easy for the child’s reflection to be lost to noise.

Advances in semiconductor technology and radar design are improving radar’s dynamic range. For example, at CES 2024, Mobileye showed an example of a radar that could detect a wooden pallet next to a metal railing at nearly 240 meters.

Software

Horsepower is becoming democratized. Scaling an electric motor to double or triple the horsepower is much cheaper than doubling the output of an internal-combustion engine. Cars that would put the supercars of yesteryear to shame are now entering the market at less than US$50,000. In combination with this, electric cars do not possess the differentiating characteristics of engines. They do not have turbo lag or differing exhaust notes; most do not even have more than one gear. In short, future cars will be differentiable not by their performance but by their software.

Software-defined vehicles, connectivity and autonomy also provide a route to perfect safety. The problem with sensors, even the best and most advanced sensors, is that they cannot see around corners or through buildings. No matter how many radar, LiDAR or camera units are fitted to a car, they cannot see that a speeding car on an adjacent city street is about to jump a red light. But connected vehicles could.

Connectivity and software offer a pathway to collective perception, in which cars and infrastructure would share their environmental knowledge with each other. Armed with such information, a connected car could react appropriately to the location and movement of every pedestrian, cyclist and vehicle on the road around it, including those beyond its line of sight. Combine collective perception with a responsible autonomous driver, and a country full of such cars could end road traffic accidents altogether.

This scenario is, of course, a long way off; autonomy on the roads is in the early days, and collective perception is a university research project. However, autonomous vehicles are already starting to show their potential safety benefits.

Safety

In October 2023, the California DMV revoked Cruise’s driverless testing permit, preventing it from operating its robotaxi program and testing cars on public roads in California. The stated reason for this was that Cruise “misrepresented” the safety of the technology. It is clear from this incident and others that AV safety claims will always and should always be subject to public scrutiny. This is not to disparage AVs or their safety, however. Rather, public oversight should be undertaken not with the motivation to eliminate autonomous cars but with the intent to improve them.

Unlike accidents involving human drivers, the intelligence gathered on a single AV collision can generate improvements that can be rolled out via over-the-air updates to the entire fleet. In this case, a collision provides an opportunity to prevent that type of collision from ever happening again, something far beyond the capabilities of human drivers.

By all metrics, the safety of AVs is improving. The data from California shows as much. Going back to 2015, when records were first kept, autonomous cars would barely make it 1,000 miles without the safety driver behind the wheel intervening. Fast forward to 2023, and Cruise submitted 576,000 miles worth of driving without a single disengagement. By comparison, IDTechEx estimates that human drivers in the U.S. are involved in collisions approximately once per 200,000 miles, and in cities like San Francisco, this rate nearly doubles. So perhaps the argument can be made that autonomous cars are already safer than human drivers.

On the other hand, a look through the data on driverless autonomous cars reveals that, on average, a driverless Waymo is involved in collisions once every 52,000 miles. Cruise fares slightly better, with one collision per 63,000 miles. So, they are colliding at roughly twice the rate of average San Franciscan drivers. But again, this is not the whole story. Out of 33 recorded collisions involving a driverless Cruise vehicle, blame can be attributed to the autonomous driving system in only six cases, which equates to once per 344,000 miles. Likewise, Waymo’s autonomous driving system could be held responsible for roughly five collisions, equating to one per 238,000 miles. So by this metric, autonomous drivers are once again safer than the average human driver.

Another metric that should be considered is the rate at which fatalities and serious injuries occur with AVs. Once again, this is lower for AVs than for human drivers.

However, there is a big issue with how this discussion is framed, and it concerns the “average” human driver. Clearly, human driving proficiency is a spectrum, encompassing those who are incapable of staying below the speed limit and off their phones, the super-vigilant types who have never had an accident and everyone in between. The question, then, is with which human drivers an autonomous car should be compared. Are autonomous drivers safe enough when they are safer than 90% of human drivers? Should they be better than every human driver? But how would that be defined? A 17-year-old with a fresh license may never have crashed, i.e., one collision per infinity miles.

Complicating matters is that the true safety record for human driving is not fully known. Not all collisions are reported, and not all dangerous situations or near-misses end in collisions. The industry just about has a grasp of how safe the average person is by measuring the rate at which people crash, cause injuries and cause fatalities. However, insufficient data exists to understand human driving safety across the full spectrum of abilities to the same depth as autonomous driving safety.

It was important to show with data and numbers that AVs could exceed the driving standards of the average human driver. However, to go beyond this and show that autonomous cars are safer than any human driver, the benchmark will need to change or evolve. Even beyond near-misses, there are many metrics that could be used to better understand human driving safety, such as how often drivers go through red lights, how often they tailgate, how often they disobey traffic rules, how often they exceed speed limits and many more. With these metrics, the industry could better understand human driving performance and make solid comparisons to build a case that AVs are safer.

Between sensors, software and safety, it is an exciting time for the automotive industry. The advent of AVs in recent years is an era-defining moment. In 100 years’ time, examples of today’s autonomous cars and robotaxis will be exhibited in museums as part of the story of how the automotive industry followed a pathway to infallible road safety.