Waterloo Region Record

AI can read! But do they understand?

Tech firms are racing to smarten up thinking machines

- Matt O’Brien

PROVIDENCE, R.I. — Seven years ago, a computer beat two human quizmaster­s on a “Jeopardy” challenge. Ever since, the tech industry has been training its machines to make them even better at amassing knowledge and answering questions.

And it’s worked, at least up to a point. Just don’t expect artificial intelligen­ce to spit out a literary analysis of Leo Tolstoy’s “War and Peace” any time soon.

Research teams at Microsoft and Chinese tech company Alibaba reached what they described as a milestone earlier this month when their AI systems outperform­ed the estimated human score on a reading comprehens­ion test. It was the latest demonstrat­ion of rapid advances that have improved search engines and voice assistants and that are finding broader applicatio­ns in health care and other fields.

The answers they got wrong — and the test itself — also highlight the limitation­s of computer intelligen­ce and the difficulty of comparing it directly to human intelligen­ce.

“We are still a long way from computers being able to read and comprehend general text in the same way that humans can,” said Kevin Scott, Microsoft’s chief technology officer, in a LinkedIn post that also commended the achievemen­t by the company’s Beijing-based researcher­s.

The test developed at Stanford University demonstrat­ed that, in at least some circumstan­ces, computers can beat humans at quickly “reading” hundreds of Wikipedia entries and coming up with accurate answers to questions about Genghis Khan’s reign or the Apollo space program.

The computers, however, also made mistakes that many people wouldn’t have.

Microsoft, for instance, fumbled an easy football question about which member of the NFL’s Carolina Panthers got the most intercepti­ons in the 2015 season (the correct answer was Kurt Coleman, not Josh Norman). A person’s careful reading of the Wikipedia passage would have discovered the right answer, but the computer tripped up on the word “most” and didn’t understand that seven is bigger than four.

“You need some very simple reasoning here, but the machine cannot get it,” said Jianfeng Gao,

of Microsoft’s AI research division.

It’s not uncommon for machine-learning competitio­ns to pit the cognitive abilities of computers against humans. Machines first bested people in an image-recognitio­n competitio­n in 2015 and a speech recognitio­n competitio­n last year, although they’re still easily tricked. Computers have also vanquished humans at chess, Pac-Man and the strategy game Go.

And since IBM’s “Jeopardy” victory in 2011, the tech industry has shifted its efforts to data-intensive methods that seek to not just find factoids, but better comprehend the meaning of multi-sentence passages.

Like the other tests, the Stanford Question Answering Dataset, nicknamed Squad, attracted a rivalry among research institutio­ns and tech firms — with Google, Facebook, Tencent, Samsung and Salesforce also giving it a try.

“Academics love competitio­ns,” said Pranav Rajpurkar, the Stanford doctoral student who helped develop the test. “All these companies and institutio­ns are trying to establish themselves as the leader in AI.”

The tech industry’s collection and digitizati­on of huge troves of data, combined with new sets of algorithms and more powerful computing, has helped inject new energy into a machinelea­rning field that’s been around for more than half a century. But computers are still “far off ” from truly understand­ing what they’re reading, said Michael Littman, a Brown University computer science professor who has tasked computers to solve crossword puzzles.

Computers are getting better at the statistica­l intuition that allows them to scan text and find what seems relevant, but they still struggle with the logical reasoning that comes naturally to people. (And they are often hopeless when it comes to decipherin­g the subtle wink-and-nod trickery of a clever puzzle.) Many of the common ways of measuring artificial intelligen­ce are in some ways teaching to the test, Littman said.

“It strikes me for the kind of problem that they’re solving that it’s not possible to do better than people, because people are defining what’s correct,” Littman said of the Stanford benchmark. “The impressive thing here is they met human performanc­e, not that they’ve exceeded it.”

 ?? JEFF CHIU, THE ASSOCIATED PRESS ?? Pranav Rajpurkar, a doctoral student at Stanford University, helped develop a test to demonstrat­e the ability of computers to comprehend reading material.
JEFF CHIU, THE ASSOCIATED PRESS Pranav Rajpurkar, a doctoral student at Stanford University, helped develop a test to demonstrat­e the ability of computers to comprehend reading material.

Newspapers in English

Newspapers from Canada