San Francisco Chronicle

Beyond board games : Artificial intelligen­ce needs a new challenge

-

When a person’s intelligen­ce is tested, there are exams. IQ tests, general knowledge quizzes, SATs.

When artificial intelligen­ce is tested, there are games. Checkers, chess, Go.

But what happens when computer programs beat humans at all of those games? This is the question AI experts must ask after a Google-developed program called AlphaGo defeated a world champion Go player in four out of five matches in a series that concluded last week.

Long a yardstick for advances in AI, the era of board game testing has come to an end, said Murray Campbell, an IBM research scientist who was part of the team that developed Deep Blue, the first computer program to beat a world chess champion.

“Games are fun and they’re easy to measure,” said Campbell. “It’s clear who won and who lost, and you always have the human benchmark,” he said. “Can you do better than a human?”

For checkers, chess and now Go, it seems the answer is now a resounding yes. Computer algorithms beat world champion-level human players in checkers and chess in the 1990s.

Go — an ancient board game developed in China that is more complex than chess — was seen as one of the last hurdles.

Board games, Campbell said, were perfect tests because they have clear rules and nothing is hidden from players. The real world is much messier and full of unknowns. What’s next, it seems, is for AI to get messy.

With AI having conquered what experts call “complete informatio­n” games — the kind in which players can see what their opponents are doing — Tuomas Sandholm, a professor at Carnegie Mellon University who studies artificial intelligen­ce, said the next step is “incomplete informatio­n games” like poker.

“The game of twoplayer-limit Texas Hold’em poker has almost been solved,” said Sandholm, who described “solving” a game as finding the optimal way of playing it. “In the larger game of two-player nolimit Texas hold ‘em poker, we’re right at the cusp of it. We currently have the world’s best computer program, but we are still not better than the very best dozen or so humans.”

Games are typically chosen for the specific challenges that researcher­s want their AI to be able to overcome. With board games such as chess and Go, computer programs are put through the ringer to see if they can learn from past matches and determine the best next move. With Texas Hold’em, it’s about interpreti­ng actions as signals, and figuring out the next best move without knowing what the opponent has.

All these things have real-world applicatio­ns, Sandholm said. In complete informatio­n games, AI can help people search through large databases and do calculatio­ns and modeling. In incomplete informatio­n games, it can be used in situations where there are lots of unknown factors, such as negotiatio­ns, cybersecur­ity and auctions, and even in planning medical treatment.

Some robotics experts believe AI will one day get messier than that, taking algorithms out of controlled game environmen­ts and into the open world.

“That’s next,” said Thomas Johnson, founder of MotionFigu­res, a startup bringing robotics and AI to toys. “The challenge will be putting AI algorithms into practice in open environmen­ts where the rules are not all given to it up-front, and adaptation is required to be successful.”

Johnson points to DARPA challenges — competitio­ns run by the U.S. Department of Defense — as an example of robots and AI being put to the test. Past DARPA challenges forced researcher­s to build robots that can walk up and down hills; many fell over or staggered like toddlers learning to walk.

“We take for granted things like balance and vision,” he said. “But for a robot, to walk up and down hills requires so many complicate­d deci- sions to be made in real time, and it’s really difficult to do.”

There’s still more that can be done in controlled environmen­ts, though.

Algorithms such as AlphaGo only know how to play Go. Oren Etzioni, executive director of the Allen Institute for Artificial Intelligen­ce, believes the next step could be for AI to learn to play (and beat world champions) at any game. Or, as his institute is doing, putting AI through standard tests like the SATs or eighth-grade science tests.

“The scientific part of it isn’t complicate­d,” Etzi- oni said.

After all, it’s not hard to get a computer program to remember and regurgitat­e facts. What is hard is getting computers to apply their knowledge to everyday situations.

“The question in the test doesn’t require the computer to give a definition of gravity or recite an equation, but to describe a real world situation,” he said. “For example, ‘There’s a ball rolling down a hill.’ This is the paradox: The hard part for the machine is easy for the human. The machine is struggling to figure out what does it mean when it says, ‘The ball is rolling down the hill’?”

There are no shortages of tests that come after Go. And that doesn’t even get into benchmarks for different types of artificial intelligen­ce, such as emotionall­y intelligen­t AI, speech-recognizin­g robots, or computers designed to understand language.

“It’s an exciting field to be in,” said Johnson. “I think it’s incredible and it makes me stop and think what the next 10, 20 years is gonna bring in AI.”

 ?? Photos by Ahn Young-joon / Associated Press ??
Photos by Ahn Young-joon / Associated Press
 ??  ?? Students watch a livestream broadcast of the match between Google’s AI program and profession­al Go player Lee Sedol in South Korea last week. The match between AI program AlphaGo and Lee is seen on TV screens in a Seoul electronic­s store.
Students watch a livestream broadcast of the match between Google’s AI program and profession­al Go player Lee Sedol in South Korea last week. The match between AI program AlphaGo and Lee is seen on TV screens in a Seoul electronic­s store.

Newspapers in English

Newspapers from United States