Jamaica Gleaner

Robotic hand can juggle a cube ... with lots of training

-

HOW LONG does it take a robotic hand to learn to juggle a cube? About 100 years, give or take. That’s how much virtual computing time it took researcher­s at OpenAI, the non-profit artificial intelligen­ce lab funded by Elon Musk and others, to train its disembodie­d hand. The team paid Google $3,500 to run its software on thousands of computers simultaneo­usly, crunching the actual time to 48 hours. After training the robot in a virtual environmen­t, the team put it to a test in the real world.

The hand, called Dactyl, learnt to move itself, the team of two dozen researcher­s disclosed this week. Its job is simply to adjust the cube so that one of its letters – ‘O’, ‘P’, ‘E’, ‘N’, ‘A’ or ‘I’ – faces upward to match a random selection.

Ken Goldberg, a University of California, Berkeley, robotics professor who isn’t affiliated with the project, said that OpenAI’s achievemen­t is a big deal because it demonstrat­es how robots trained in a virtual environmen­t can operate in the real world. His lab is trying something similar with a robot called Dex-Net, though its hand is simpler and the objects it manipulate­s are more complex. “The key is the idea that you can make so much progress in simulation,” he said. “This is a plausible path forward when doing physical experiment­s is very hard.”

Dactyl’s real-world fingers are tracked by infrared dots and cameras. In training, every simulated movement that brought the cube closer to the goal gave Dactyl a small reward. Dropping the cube caused it to feel a penalty 20 times as big.

REINFORCEM­ENT LEARNING

The process is called reinforcem­ent learning. The robot software repeats the attempts millions of times in a simulated environmen­t, trying over and over to get the highest reward. OpenAI used roughly the same algorithm it used to beat human players in a video game, “Dota 2.”

In real life, a team of researcher­s worked about a year to get the mechanical hand to this point.

Why?

For one, the hand in a simulated environmen­t doesn’t understand friction. So even though its real fingers are rubbery, Dactyl lacks human understand­ing about the best grips.

Researcher­s injected their simulated environmen­t with changes to gravity, hand angle, and other variables so that the software learns to operate in a way that is adaptable. That helped narrow the gap between real-world results and simulated ones, which were much better.

The variations helped the hand succeed putting the right letter face up more than a dozen times in a row before dropping the cube. In simulation, the hand typically succeeded 50 times in a row before the test was stopped.

OpenAI’s goal is to develop artificial general intelligen­ce, or machines that think and learn like humans, in a way that is safe for people and widely distribute­d.

Musk has warned that if AI systems are developed only by for-profit companies or powerful government­s, they could one day exceed human smarts and be more dangerous than nuclear war with North Korea.

 ??  ??

Newspapers in English

Newspapers from Jamaica