The Standard (St. Catharines)

AI has a dirty little secret: It’s powered by real people

- RYAN NAKASHIMA

There’s a dirty little secret about artificial intelligen­ce: It’s powered by an army of real people.

From makeup artists in Venezuela to women in conservati­ve parts of India, people around the world are doing the digital equivalent of needlework — drawing boxes around cars in street photos, tagging images, and transcribi­ng snatches of speech that computers can’t quite make out.

Such data feeds directly into “machine learning” algorithms that help self-driving cars wind through traffic and let Alexa figure out that you want the lights on. Many such technologi­es wouldn’t work without massive quantities of this human-labeled data.

These repetitive tasks pay pennies apiece. But in bulk, this work can offer a decent wage in many parts of the world — even in the U.S. This burgeoning but largely unseen cottage industry represents the foundation of a technology that could change humanity forever: AI that will drive us around, execute verbal commands without flaw, and, possibly, one day think on its own. This human input industry has long been nurtured by search engines Google and Bing, who for more than a decade have used people to rate the accuracy of their results. Since 2005, Amazon’s Mechanical Turk service, which matches freelance workers with temporary online jobs, has also made crowdsourc­ed data entry available to researcher­s worldwide. More recently, investors have poured tens of millions of dollars into startups like Mighty AI and CrowdFlowe­r, which are developing software that makes it easier to label photos and other data, even on smartphone­s.

Venture capitalist S. “Soma” Somasegar says he sees “billions of dollars of opportunit­y” in servicing the needs of machine learning algorithms. His firm, Madrona Venture

Group, invested in Mighty AI. Humans will be in the loop “for a long, long, long time to come,” he says. Accurate labelling could make the difference between a self-driving car distinguis­hing between the sky and the side of a truck — a distinctio­n Tesla’s Model S failed in the first known fatality involving self-driving systems in 2016.

“We’re not building a system to play a game, we’re building a system to save lives,” says Mighty AI CEO Daryn Nakhuda.

Marjorie Aguilar, a 31-year-old freelance makeup artist in Maracaibo, Venezuela, spends four to six hours a day drawing boxes around traffic objects to help train self-driving systems for Mighty AI. She earns about 50 cents an hour, but in a crisis-wracked country with runaway inflation, just a few hours’ work can pay a month’s rent in bolivars.

“It doesn’t sound like a lot of money, but for me it’s pretty decent,” she says. “You can imagine how important it is for me getting paid in U.S. dollars.”

Aria Khrisna, a 36-year-old father of three in Tegal, Indonesia, says doing things like adding word tags to clothing pictures on websites such as eBay and Amazon pays him about $100 a month, roughly half his income. And for 25-year-old Shamima Khatoon, her job annotating cars, lane markers and traffic lights at an all-female outpost of data-labeling company iMerit in Metiabruz, India, represents the only chance she has to work outside the home in her conservati­ve Muslim community.

Major automakers like Toyota, Nissan and Ford, ride-hailing companies like Uber and other tech giants like Alphabet Inc.’s Waymo are paying reams of labelers, often through third-party vendors. The benefits of greater accuracy can be immediate.

At InterConti­nental Hotels Group, every call that its digital assistant Amelia can take from a human saves $5 to $10, says informatio­n technology director Scot Whigham.

When Amelia fails, the program listens while a call is rerouted to one of about 60 service desk workers. It learns from their response and tries the technique out on the next call, freeing up human employees to do other things.

“We’ve transforme­d those jobs,” Whigham says.

When a computer can’t make out a customer call to the Hyatt Hotels chain, an audio snippet is sent to AI-powered call centre Interactio­ns in an old brick building in Franklin, Mass.

There, while the customer waits on the phone, one of a roomful of headphone-wearing “intent analysts” transcribe­s everything from misheard numbers to profanitie­s and quickly directs the computer how to respond. That informatio­n feeds back into the system. “Next time through, we’ve got a better chance of being successful,” says Robert Nagle, Interactio­ns’ chief technology officer.

Researcher­s have tried to find workaround­s to human-labeled data, but the results are often inadequate. In a project that used Google Street View images of parked cars to estimate the demographi­c makeup of neighbourh­oods, then-Stanford researcher Timnit Gebru tried to train her AI by scraping Craigslist photos of cars for sale that were labelled by their owners.

But the product shots didn’t look anything like the car images in Street View, and the program couldn’t recognize them. In the end, she says, she spent $35,000 to hire auto dealer experts to label her data.

The need for human labelers is “enormous” and “dynamic,” says Robin Bordoli, CEO of labelling technology company CrowdFlowe­r. “You can’t trust the algorithm 100 per cent.” At the moment, figuring out how to get computers to learn without so-called “ground truth” data provided by humans remains an open research question.

Trevor Darrell, a machine learning expert at the University of California Berkeley, says he expects it will be five to 10 years before computer algorithms can learn to perform without the need for human labelling.

His group alone spends hundreds of thousands of dollars a year paying people to annotate images. “Right now, if you’re selling a product and you want perfection, it would be negligent not to invest the money in that kind of annotation,” he says.

 ??  ?? STEVEN SENNE Jessica McShane, an employee at Interactio­ns Corp., monitors person-to-computer communicat­ions, helping computers understand what a human is saying, in the "intent analysis" room at the company's headquarte­rs in Franklin, Mass.
STEVEN SENNE Jessica McShane, an employee at Interactio­ns Corp., monitors person-to-computer communicat­ions, helping computers understand what a human is saying, in the "intent analysis" room at the company's headquarte­rs in Franklin, Mass.

Newspapers in English

Newspapers from Canada