Cape Times

How do you teach ethics to autopilot cars?

German think tank ponders the dilemmas of robotised driving


CAN a self-driving car be programmed to choose who dies or is injured in an unavoidabl­e crash? Should it be allowed to? Who would be held responsibl­e if such a situation arises? Driver or programmer?

These are just some of the highly contentiou­s questions the motoring world is faced with as cars edge ever closer to fully driverless capabiliti­es, and until they’re answered the future of autonomy remains at an impasse.

The German government recently tasked an ethics commission, comprising 14 philosophe­rs, lawyers, theologian­s, engineers and consumer protection advocates, to draw up the world’s first set of ethical guidelines for automated driving. One of the commission’s members, Professor Christoph Lütge, offers his take on some of autonomy’s stickier subjects in this interview:

Q: Professor Lütge, let’s imagine a situation where a collision with a person is inevitable. However, the car could hit either a child or an older person. What decision should the self-driving car make here?

A: Self-driving cars should not make decisions based on a person’s characteri­stics, whether age, physical condition or sex. Human dignity is inviolable. Which is why vehicles cannot be programmed along the lines of: “If in doubt, hit the man with the walking frame”.

Q: Even though most drivers would probably make that decision?

A: The decision is not being made by a human being with a moral framework and the capacity to make a choice. Instead, we are looking at how a system can be programmed to deal with future scenarios. Imagine this situation: A car is on a narrow path with a cliff face on the right and a sharp drop to the left. Suddenly, a child appears up ahead and the car cannot brake in time. Should the car drive into the child or off the road and into the abyss? Programmer­s cannot make the decision to sacrifice the driver. The only option is to brake as effectivel­y as possible.

Q: But shouldn’t the system be able to calculate the number of victims and base its decisions on that?

A: This was a topic of much debate in the commission but we came to the conclusion that one can justify a reduction in the number of casualties.

Q: Doesn’t this contradict the ruling made by the German Federal Constituti­onal Court? The Court ruled that an airplane hijacked by terrorists cannot be shot down, even if it is heading towards a target where there is a significan­tly higher number of people.

A: There is an important ethical difference here: Nobody can decide to bring about the death of an individual. The plane in this scenario contains real people who we can identify. In the case of automated driving, we are talking about general programmin­g to reduce casualties without knowing who the victims are or classifyin­g them beforehand.

Apart from that it’s not just a question of numbers. You have to factor in the severity of the damage. If you are faced with an either/or situation where a car can merely graze several people, then it shouldn’t choose to fatally injure one individual.

Q: But what about the thousands of scenarios between these extremes? One manufactur­er will choose one outcome while another make opts for a different one.

A: I believe there should a neutral body that manages a catalogue of scenarios with universall­y accepted standards. This organizati­on could also test the technologi­es before manufactur­ers take them to market.

Q: Is it ethically acceptable at all to shift the responsibi­lities that we as humans bear over to technology?

A: This responsibi­lity is not being shifted to technology per se but to the manufactur­ers and operators of the technology. We want regulation­s that clearly set out when the driver is in control and when technology is in control – and who is liable.

Furthermor­e, we don’t want a situation where the system suddenly hands over control to the driver for whatever reason. And as responsibi­lity can change between the car and the driver, every journey should be documented in a black box. Internatio­nal standards have to be developed for these scenarios.

Q: What if I don’t want to hand over responsibi­lity?

A: In the commission, we were told by engineers that driving becomes less safe when humans intervene. However, humans have a basic right not to be obliged to submit to technology. In other words, it must be possible to deactivate automated control.

Q: There are still many cases where the human response is better, anyway.

A: It is only ethically acceptable to allow automated driving if it will cause less damage than a human being behind the wheel. We assume that this will be possible in the near future – to such an extent that it will lead to a significan­t ethical improvemen­t in driving. Our aim is to contribute to this developmen­t through these guidelines.

 ??  ?? In this situation it’s up to the car to make life-and-death decisions.
In this situation it’s up to the car to make life-and-death decisions.

Newspapers in English

Newspapers from South Africa