The everyday ethical challenges of self-driving cars

The biggest ethical challenges for self-driving cars arise in mundane situations, not when crashes are unavoidable

Johannes Himmelreich
Wednesday 25 April 2018 05:15 EDT
Comments
Extreme crash scenarios are a distraction away from important everyday issues
Extreme crash scenarios are a distraction away from important everyday issues (Grendelkhan)

Your support helps us to tell the story

From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.

At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.

The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.

Your support makes all the difference.

A lot of discussion and ethical thought about self-driving cars have focused on tragic dilemmas, like hypotheticals in which a car has to decide whether to run over a group of schoolchildren or plunge off a cliff, killing its own occupants. But those sorts of situations are extreme cases.

As the most recent crash – in which a self-driving car killed a pedestrian in Tempe, Arizona – demonstrates, the mundane, everyday situations at every pedestrian crossing, turn and intersection present much harder and broader ethical quandaries.

Ethics of extremes

As a philosopher working with engineers in Stanford’s Centre for Automotive Research, I was initially surprised that we spent our lab meetings discussing what I thought was an easy question: how should a self-driving car approach a pedestrian crossing?

My assumption had been that we would think about how a car should decide between the lives of its passengers and the lives of pedestrians. I knew how to think about such dilemmas because these crash scenarios resemble a famous philosophical brainteaser called the “trolley problem”. Imagine a runaway trolley is hurling down the tracks and is bound to hit either a group of five or a single person – would you kill one to save five?

However, many philosophers nowadays doubt that investigating such questions is a fruitful avenue of research. Barbara Fried, a colleague at Stanford, for example, has argued that tragic dilemmas make people believe ethical quandaries mostly arise in extreme and dire circumstances.

In fact, ethical quandaries are ubiquitous. Everyday, mundane situations are surprisingly messy and complex, often in subtle ways. For example, should your city spend money on a diabetes prevention programme or on more social workers? Should your local department of public health hire another inspector for restaurant hygiene standards, or continue a programme providing free needles and injection supplies?

(xkcd)

These questions are extremely difficult to answer because of uncertainties about the consequences – such as who will be affected and to what degree. The solutions philosophers have proposed for extreme and desperate situations are of little help here.

The problem is similar with self-driving cars. Thinking through extreme situations and crash scenarios cannot help answer questions that arise in mundane situations.

A challenge at pedestrian crossings

One could ask, what can be so hard about mundane traffic situations like approaching a pedestrian crossing, driving through an intersection, or making a left turn? Even if visibility at the pedestrian crossing is limited and it is sometimes hard to tell whether a nearby pedestrian actually wants to cross the street, drivers cope with this every day.

But for self-driving cars, such mundane situations pose a challenge in two ways.

First, there is the fact that what is easy for humans is often hard for machines. Whether it is recognising faces or riding bicycles, we are good at perception and mechanical tasks because evolution built these skills for us. That, however, makes these skills hard to teach or engineer. This is known as “Moravec’s paradox.”

Second, in a future where all cars are self-driving, small changes to driving behaviour would make a big difference in the aggregate. Decisions made by engineers today, in other words, will determine not how one car drives but how all cars drive. Algorithms become policy.

Engineers teach computers how to recognise faces and objects using methods of machine learning. They can use machine learning also to help self-driving cars imitate how humans drive. But this isn’t a solution, it doesn’t solve the problem that wide-ranging decisions about safety and mobility are made by engineers.

Furthermore, self-driving cars shouldn’t drive like people. Humans aren’t actually very good drivers. And they drive in ethically troubling ways, deciding whether to yield at pedestrian crossings, based on pedestrians’ age, race and income. For example, researchers in Portland have found that black pedestrians are passed by twice as many cars and had to wait a third longer than white pedestrians before they can cross.

Self-driving cars should drive more safely, and more fairly than people do.

Mundane ethics

The ethical problems deepen when you attend to the conflicts of interest that surface in mundane situations such as pedestrian crossings, turns and intersections.

For example, the design of self-driving cars needs to balance the safety of others – pedestrians or cyclists – with the interests of cars’ passengers. As soon as a car goes faster than walking pace, it is unable to prevent from crashing into a child that might run onto the road in the last second. But walking pace is, of course, way too slow. Everyone needs to get to places. So how should engineers strike the balance between safety and mobility? And what speed is safe enough?

There are other ethical questions that come up as well. Engineers need to make trade-offs between mobility and environmental impacts. When they’re applied across all the cars in the country, small changes in computer-controlled acceleration, cornering and braking can have huge effects on energy use and pollution emissions. How should engineers trade off travel efficiency with environmental impact?

What should the future of traffic be?

Mundane situations pose novel engineering and ethical problems, but they also lead people to question basic assumptions of the traffic system.

For myself, I began to question whether we need places called “pedestrian crossings” at all? After all, self-driving cars can potentially make it safe to cross a road anywhere.

And it is not only pedestrian crossings that become unnecessary. Traffic lights at intersections could be a thing of the past as well. Humans need traffic lights to make sure everyone gets to cross the intersection without crash and chaos. But self-driving cars could coordinate among themselves smoothly.

The bigger question here is this: given that self-driving cars are better than human drivers, why should the cars be subject to rules that were designed for human fallibility and human errors? And to extend this thought experiment, consider also the more general question: if we, as a society, could design our traffic system from scratch, what would we want it to look like?

Because these hard questions concern everyone in a city or in a society, they require a city or society to agree on answers. That means balancing competing interests in a way that works for everybody – whether people think only about pedestrian crossings or about the traffic system as a whole.

With self-driving cars, societies can redesign their traffic systems. From the pedestrian crossings to overall traffic design – it is mundane situations that raise really hard questions. Extreme situations are a distraction.

The trolley problem does not answer these hard questions.

Johannes Himmelreich is an interdisciplinary ethics fellow at Stanford University McCoy Family Centre for Ethics in Society. This article was first published in The Conversation (theconversation.com)

Join our commenting forum

Join thought-provoking conversations, follow other Independent readers and see their replies

Comments

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in