The National - News

HOW ACCENT BIAS IN VOICE TECHNOLOGY CREATES AN UNFAIR WORLD

▶ As the system is added to more devices, Rhodri Marsden explores its skewed nature towards English of highly educated white Americans

-

Anyone who has used a voice assistant such as Apple’s Siri or Amazon’s Alexa will have occasional­ly struggled to make themselves understood. Perhaps the device plays the wrong music, or puts unusual items on a shopping list, or emits a plaintive “didn’t quite catch that”. But for people who speak with an accent, these devices can be unusable.

The inability of speech recognitio­n systems to understand accents found in Scotland, Turkey, the southern states of the US or any number of other places is widely documented, and yet the problem persists. With uses of the technology now spreading beyond the domestic, academics are warning that biased systems could lead to new forms of discrimina­tion, purely because of someone’s accent.

“It’s one of the questions that you don’t see big tech responding to,” says Halcyon Lawrence a professor of technical communicat­ion at Towson University in Maryland who is from Trinidad and Tobago. “There’s never a statement put out. There’s never a plan that’s articulate­d. And that’s because it’s not a problem for big tech. But it’s a problem for me, and large groups of people like me.”

Speech recognitio­n systems can only recognise accents they’ve been trained to understand. To learn how to interpret the accent of someone from Trinidad, Eswatini or the UAE, a system needs voice data, along with an accurate transcript­ion of that data, which inevitably has to be done by a human being. It’s a painstakin­g and expensive process to demonstrat­e to a machine what a particular word sounds like when it’s spoken by a particular community, and perhaps inevitably, existing data is heavily skewed towards English as typically spoken by white, highly educated Americans.

A study called Racial Disparitie­s in Automated Speech Recognitio­n, published last year by researcher­s at Stanford University, illustrate­s the stark nature of the problem. It analysed systems developed by Amazon, Apple, Google, IBM and Microsoft, and found that in every case the error rates for black speakers were nearly double that of white people. In addition, it found that the errors were not caused by grammar, but by “phonologic­al, phonetic, or prosodic characteri­stics”; in other words, accent.

Allison Koenecke, who led the study, believes that a two-fold improvemen­t in the system is needed. “It needs resources to ethically collect data and ensure that the people working on these products are also diverse,” she says. “While tech companies may have the funds, they may not have known that they needed to prioritise this issue before external researcher­s shone a light on it.”

Lawrence, however, believes that the failings are no accident. “What, for me, shows big tech’s intention is when they decide to release a new accent to the market and where that is targeted,” she says.

“If you plot it on a map, you can’t help but notice that the Global South is not a considerat­ion, despite the numbers of English speakers there. So you begin to see that this is an economic decision.”

Arabic also poses a particular challenge – not simply because of the many sub-dialects, but inherent difficulti­es such as the lack of capital letters, recognisin­g proper nouns and predicting a word’s vowels based on context. Substantia­l resources are being thrown at this problem, but the current situation is the same as with English: large communitie­s technologi­cally disenfranc­hised.

Why is this of particular concern? Beyond the world of smart speakers lies a much bigger picture. “There are many higher-stakes applicatio­ns with much worse consequenc­es if the underlying technologi­es are biased,” says Koenecke. “One example is court transcript­ions, where court reporters are starting to use automatic speech recognitio­n technologi­es. If they aren’t accurate at transcribi­ng cases, you have obvious repercussi­ons.”

Lawrence is particular­ly concerned about the way people drop their accent in order to be understood, rather than the technology working harder to understand them. “There’s an expectatio­n that we adapt our accent, and that’s what gets replicated in the device. It would not be an acceptable demand on somebody to change the colour of their skin, so why is it acceptable to demand we change our accents?,” she says.

Money, as ever, lies at the root of the problem. Lawrence believes strongly that big tech has to be urged to look beyond its profit margin. “It’s one of the reasons why I believe that we’re going to see more and more smaller independen­t developers do this kind of work,” she says.

One of those developers, a British company called Speechmati­cs, is at the forefront, using what it calls “self-supervised learning” to introduce its speech recognitio­n systems to a new world of voices. “We’re training on over a million hours of unlabelled audio, and constructi­ng systems that can learn interestin­g things, autonomous­ly run,” says Will Williams, vice president of machine learning at Speechmati­cs.

The crucial point: this is voice data that hasn’t been transcribe­d. “If you have the right kind of diversity of data, it will learn to generalise across voices, latch on quickly and understand what’s going on.” Using datasets from the Stanford study, Speechmati­cs has already reported a 45 per cent reduction in errors when using its system.

An organisati­on called ML Commons, which has Google and Microsoft as two of its more than 50 founding members, is now looking for new ways to create systems that are accent-agnostic.

It’s a long road ahead, but Koenecke is optimistic. “Hopefully, as different speechto-text companies decide to invest in more diverse data and more diverse teams of engineers and product managers, we will see something that reflects more closely what we see in real life.”

It’s one of the questions that you don’t see big tech responding to

... because it’s not a problem for them HALCYON LAWRENCE Professor of tech communicat­ions

 ?? Getty ??
Getty

Newspapers in English

Newspapers from United Arab Emirates