Qatar Tribune

Self-Driving Cars May Have A Racism Problem

- THEODORE KIM (Theodore Kim is an associate professor of computer science at Yale University.)

TESLA recently announced the newest version of its self-driving car software, following the software’s role in a dozen reported collisions with emergency vehicles that are the subject of a federal agency probe. While these collisions happened for a variety of reasons, a major factor may be that the artificial intelligen­ce driving the car is not used to seeing flashing lights and vehicles pulled over on the shoulder, so the underlying algorithms react in unpredicta­ble and catastroph­ic ways.

Modern AI systems are “trained” on massive datasets of photograph­s and video footage from various sources, and use that training to determine appropriat­e behaviour. But, if the footage doesn’t include lots of examples of specific behaviours, like how to slow down near emergency vehicles, the AI will not learn the appropriat­e behaviours. Thus, they crash into ambulances.

Given these types of disastrous failures, one recent trend in machine learning is to identify these neglected cases and create “synthetic” training data to help the AI learn. Using the same algorithms that Hollywood used to assemble the Incredible

Hulk in The Avengers: Endgame from a stream of ones and zeros, photoreali­stic images of emergency vehicles that never existed in real life are conjured from the digital ether and fed to the AI.

I have been designing and using these algorithms for the last 20 years, starting with the software used to generate the sorting hat in Harry Potter And The Sorcerer’s Stone, up through recent films from Pixar, where I used to be a senior research scientist.

Using these algorithms to train AIs is extremely dangerous, because they were specifical­ly designed to depict white humans. All the sophistica­ted physics, computer science and statistics that undergird this software were designed to realistica­lly depict the diffuse glow of pale, white skin and the smooth glints in long, straight hair.

In contrast, computer graphics researcher­s have not systematic­ally investigat­ed the shine and gloss that characteri­ses dark and Black skin, or the characteri­stics of Afro-textured hair. As a result, the physics of these visual phenomena are not encoded in the Hollywood algorithms.

To be sure, synthetic Black people have been depicted in film, such as in last year’s Pixar movie Soul. But behind the scenes, the lighting artists found that they had to push the software far outside its default settings and learn all new lighting techniques to create these characters. These tools were not designed to make non-white humans even the most technicall­y sophistica­ted artists in the world strained to use them effectivel­y.

Regardless, these same white-human generation algorithms are currently being used by start-up companies like Datagen

and Synthesis AI to generate “diverse” human datasets specifical­ly for consumptio­n by tomorrow’s AIs.

A critical examinatio­n of some of their results reveal the same patterns. White skin is faithfully depicted, but the characteri­stic shine of Black skin is either disturbing­ly missing, or distressin­gly overlighte­d.

Once the data from these flawed algorithms are ingested by AIs, the provenance of their malfunctio­ns will become nearimposs­ible to diagnose. When Tesla Roadsters start disproport­ionally running over Black paramedics, or Oakland residents with natural hairstyles, the cars won’t be able to report that “nobody told me how Black skin looks in real life”. The behaviour of artificial neural networks is notoriousl­y difficult to trace back to specific problems in their training sets, making the source of the issue extremely opaque.

Synthetic training data are a convenient shortcut when real-world collection is too expensive. But AI practition­ers should be asking themselves: Given the possible consequenc­es, is it worth it If the answer is no, they should be pushing to do things the hard way: by collecting the real-world data.

Hollywood should do its part and invest in the research and developmen­t of algorithms that are rigorously, measurably, demonstrab­ly capable of depicting the full spectrum of humanity.

Not only will it expand the range of stories that can be told, but it could literally save someone’s life. Otherwise, even though you may recognise that Black lives matter, pretty soon your car won’t.

 ?? ??

Newspapers in English

Newspapers from Qatar