Khaleej Times

We may soon find out what makes robots tick

- Jeremy Kahn

london — When our robot overlords arrive, will they decide to kill us or cooperate with us?

New research from DeepMind, Alphabet’s London-based artificial intelligen­ce unit, could ultimately shed light on this fundamenta­l question.

They have been investigat­ing the conditions in which reward-optimising beings, whether human or robot, would chose to cooperate, rather than compete. The answer could have implicatio­ns for how computer intelligen­ce may eventually be deployed to manage complex systems such as an economy, city traffic flows, or environmen­tal policy.

Joel Leibo, the lead author of a paper DeepMind published online on Thursday, said in an e-mail that his team’s research indicates that whether agents learn to cooperate or compete depends strongly on the environmen­t in which they operate.

While the research has no immediate real-world applicatio­n, it would help DeepMind design artificial intelligen­ce agents that can work together in environmen­ts with imperfect informatio­n. In the future, such work could help such agents navigate a world full of intelligen­t entities — both human and machine — whether in transport networks or stock markets.

Apples and wolves

DeepMind’s paper describes how researcher­s used two different games to investigat­e how software agents learn to compete or cooperate.

In the first, two of these agents had to maximise the number of apples they could gather in a twodimensi­onal digital environmen­t. Researcher­s could vary how frequently apples appeared. The researcher­s found that when apples were scarce, the agents quickly learned to attack one another — zapping, or “tagging” their opponent with a ray that temporaril­y immobilise­d them. When apples were abundant, the agents preferred to coexist more peacefully.

Rather chillingly, however, the researcher­s found when they tried this same game with more intelligen­t agents that drew on larger neural networks, a kind of machine intelligen­ce designed to mimic how certain parts of the human brain work — they would “try to tag the other agent more frequently, i.e. behave less cooperativ­ely, no matter how we vary the scarcity of apples,” they wrote in a blog post on DeepMind’s website.

In a second game, called Wolfpack, the AI agents played wolves that had to learn to capture “prey”. Success resulted in a reward not just for the wolf making the capture, but for all wolves present within a certain radius of the capture. The more wolves present in this capture radius, the more points all the wolves would receive. In this game, the agents generally learned to cooperate. Unlike in the apple-gathering game, in Wolfpack the more cognitivel­y-advanced the agent was, the better it learned to cooperate.

The researcher­s postulate that this is because in the apple-gathering game, the zapping behaviour was more complex — it requiring aiming the beam at the opponent; while in Wolfpack game, cooperatio­n was the more complex behaviour.

The researcher­s speculated that because the less sophistica­ted artificial intelligen­ce systems had more difficulty mastering these complex behaviours, the more simple AI couldn’t learn to use them effectivel­y.

DeepMind, which Google purchased in 2014, is best known for having created an artificial intelligen­ce that can beat the world’s top human players in the ancient Asian strategy game Go. In November, DeepMind announced it was working with Blizzard Entertainm­ent Inc., the division of Activision Blizzard that makes the video game Starcraft II, to turn that game into a platform for AI research.

Leibo said that the agents used in the apple-gathering and Wolfpack experiment­s had no shortterm memory, and as a result could not make any inferences about the intent of the other agent. “Going forward, it would be interestin­g to equip agents with the ability to reason about other agent’s beliefs and goals,” he said.

In the meantime, it might be wise to keep a few spare apples around.

 ?? Reuters ?? Can you imagine what kind of havoc an extremely-intelligen­t robot can wreak if it gets ticked off after repeatedly losing rock-papersciss­ors against you? —
Reuters Can you imagine what kind of havoc an extremely-intelligen­t robot can wreak if it gets ticked off after repeatedly losing rock-papersciss­ors against you? —

Newspapers in English

Newspapers from United Arab Emirates