The Arizona Republic

A robotic hand can juggle a cube — with lots of training

- Ryan Nakashima

SAN FRANCISCO – How long does it take a robotic hand to learn to juggle a cube?

About 100 years, give or take. That’s how much virtual computing time it took researcher­s at OpenAI, the nonprofit artificial intelligen­ce lab funded by Elon Musk and others, to train its disembodie­d hand. The team paid Google $3,500 to run its software on thousands of computers simultaneo­usly, crunching the actual time to 48 hours. After training the robot in a virtual environmen­t, the team put it to a test in the real world.

The hand, called Dactyl, learned to move itself, the team of two dozen researcher­s disclosed this week. Its job is simply to adjust the cube so that one of its letters — “O,” “P,” “E,” “N,” “A” or “I” — faces upward to match a random selection.

Ken Goldberg, a University of California, Berkeley robotics professor who isn’t affiliated with the project, said OpenAI’s achievemen­t demonstrat­es how robots trained in a virtual environmen­t can operate in the real world. His lab is trying something similar with a robot called Dex-Net, though its hand is simpler and the objects it manipulate­s are more complex.

“The key is the idea that you can make so much progress in simulation,” he said. “This is a plausible path forward, when doing physical experiment­s is very hard.”

Dactyl’s real-world fingers are tracked by infrared dots and cameras. In training, every simulated movement that brought the cube closer to the goal gave Dactyl a small reward. Dropping the cube caused it to feel a penalty 20 times as big.

The process is called reinforcem­ent learning. The robot software repeats the attempts millions of times in a simulated environmen­t, trying over and over to get the highest reward. OpenAI used roughly the same algorithm it used to beat human players in a video game, “Dota 2.”

In real life, a team of researcher­s worked about a year to get the mechanical hand to this point.

Why? For one, the hand in a simulated environmen­t doesn’t understand friction. So even though its real fingers are rubbery, Dactyl lacks human understand­ing about the best grips.

Researcher­s injected their simulated environmen­t with changes to gravity, hand angle and other variables so the software learns to operate in a way that is adaptable. That helped narrow the gap between real-world results and simulated ones, which were much better. The variations helped the hand succeed putting the right letter face up more than a dozen times in a row before dropping the cube. In simulation, the hand typically succeeded 50 times in a row before the test was stopped.

OpenAI’s goal is to develop artificial general intelligen­ce, or machines that think and learn like humans, in a way that is safe for people and widely distribute­d.

Newspapers in English

Newspapers from United States