Boston Sunday Globe

Is AI really as good as advertised?

- By Elizabeth Svoboda Elizabeth Svoboda, a writer in San Jose, Calif., is the author of “What Makes a Hero?: The Surprising Science of Selflessne­ss.”

Like a many-tentacled Kraken, new artificial intelligen­ce is set to strangle human dominion on Earth — or so some experts now believe. The so-called godfather of AI, Geoffrey Hinton, quit his post at Google out of concern about where the technology is headed. AI, he thinks, will soon manipulate humans like an adult plucking a lollipop from a toddler. “Smart things can outsmart us,” he said at this year’s Emtech Digital event at MIT. “If these things get carried away with getting more control, we’re in trouble.” Data scientist Ken Jee put the anticipate­d AI-human showdown in even starker terms: “We must adapt to survive or else face extinction.”

But these highflown what-ifs may run into a more banal reality: Many AI programs simply don’t work that well. The AI errors and hallucinat­ions we see now are hardly flukes — for years, AI has had major functional­ity problems that eager adopters have overlooked.

Though you wouldn’t know it from all the recent hype, lots of AI ventures are moonshots that never quite land. AI body scan readers have failed to replace human radiologis­ts, despite prediction­s of the specialty’s demise; and safe, dependable self-driving cars have so far proved elusive. In a 2023 Altair Engineerin­g survey, business respondent­s reported that 36 percent of their AI projects flopped. In some cases, the stats are much worse. When a University of Cambridge team studied 62 AI models in 2021 that were programmed to diagnose COVID from chest X-rays or CT scans, they found that exactly none of the models was fit for clinical use due to various flaws and biases.

Whether an AI project dies on the research vine or sputters after it’s deployed (thanks to “launch first, fix later” credos), low-quality data input is often a prime culprit.

AI systems learn by mainlining data, so if you want your AI system to, say, diagnose COVID from a chest scan, you’ll feed in as many previous scans as you can, specifying which scans came from people with COVID and which ones did not. Later, when the system reads a new scan, it will use its accumulate­d knowledge to judge whether or not COVID is present.

But in practice, says Brown University AI researcher Lizzie Kumar, bad AI training data has long been rampant, leading to bad system performanc­e. Some of the COVID diagnosis programs in the University of Cambridge study were learning from biased data, such as unusual chest scans that did not represent general population scan results — a misstep that set them up for failure.

Likewise, some AI tools use faulty criminal justice data — riddled with biases and reporting errors — in trying to predict things like whether a given criminal will reoffend, yielding skewed results that can lead judges to impose too harsh sentences. “It’s hard to make a philosophi­cally sound argument that you should be using that data,” says Kumar, who is coauthor of a 2022 paper called “The Fallacy of AI Functional­ity.” “Why would you ever trust that data as ground truth?”

While press coverage has focused on notorious cases in which, for instance, predictive AI treats Black defendants differentl­y from white ones, these systems can also fail much more broadly. The accuracy of COMPAS, a popular recidivism risk calculator, hovered just above 60 percent in one study regardless of defendants’ race. And yet many US courts have relied on it for years.

It’s tempting to assume such shortcomin­gs can be resolved with more thorough and accurate training data, but AI has shown other deficits that even the best data can’t remedy. Generative AI systems like ChatGPT make a flurry of opaque calculatio­ns to devise novel replies to questions. That process veers off the rails when chatbots make up facts, a surprising­ly difficult glitch to correct. Because programs like ChatGPT make such complex connection­s between data points to generate answers, humans often cannot understand their reasoning, a longstandi­ng AI issue known as the “black box problem.” Since we can’t see the box’s inner workings, fixing them is akin to wielding a wrench blindfolde­d. Of course, that hasn’t stopped companies from forging ahead with generative AI tools and sweet-talking us into trusting them.

To protect us from our own credulity, critics are urging government regulators to crack down on companies whose AI products don’t work reliably. The Federal Trade Commission is “already signaling that they’re on board with this principle that you should be protected from unsafe or ineffectiv­e systems,” Kumar says. But grassroots efforts could prove just as crucial. Landlords get to decide whether they’ll use AI-based tenant screening aids that claim fairness but discrimina­te against many applicants. Health insurance companies get to decide whether they’ll reimburse providers for offering patients unproven AI mental health apps. And editors get to decide whether they’ll trust AI compositio­n tools that churn out dubious facts for the masses.

Will artificial intelligen­ce someday reason circles around us, as Hinton warns? For now, that almost seems beside the point. Futurists will continue generating splashy headlines, but for the rest of us, the next best move is both simpler and more profound: Evaluate each AI system individual­ly — with the same scrutiny we’d bring to a telemarket­ing pitch or a “One Weird Trick” email — and act accordingl­y.

 ?? CARLOS OSORIO/AP ?? After being falsely accused of a carjacking when she was eight months pregnant, Porcha Woodruff is suing the City of Detroit over what she says is an overrelian­ce on facial recognitio­n technology. Facial recognitio­n is one of several AI-related technologi­es called out by researcher­s in a 2022 paper, “The Fallacy of AI Functional­ity.”
CARLOS OSORIO/AP After being falsely accused of a carjacking when she was eight months pregnant, Porcha Woodruff is suing the City of Detroit over what she says is an overrelian­ce on facial recognitio­n technology. Facial recognitio­n is one of several AI-related technologi­es called out by researcher­s in a 2022 paper, “The Fallacy of AI Functional­ity.”

Newspapers in English

Newspapers from United States