Arab Times

A robotic hand can juggle a cube — with lots of training

-

How long does it take a robotic hand to learn to juggle a cube? About 100 years, give or take. That’s how much virtual computing time it took researcher­s at OpenAI, the non-profit artificial intelligen­ce lab funded by Elon Musk and others, to train its disembodie­d hand. The team paid Google $3,500 to run its software on thousands of computers simultaneo­usly, crunching the actual time to 48 hours. After training the robot in a virtual environmen­t, the team put it to a test in the real world.

The hand, called Dactyl, learned to move itself, the team of two dozen researcher­s disclosed this week. Its job is simply to adjust the cube so that one of its letters — “O,” “P,” “E,” “N,” “A” or “I’’ — faces upward to match a random selection.

Ken Goldberg, a University of California, Berkeley robotics professor who isn’t affiliated with the project, said OpenAI’s achievemen­t is a big deal because it demonstrat­es how robots trained in a virtual environmen­t can operate in the real world. His lab is trying something similar with a robot called Dex-Net, though its hand is simpler and the objects it manipulate­s are more complex.

“The key is the idea that you can make so much progress in simulation,” he said. “This is a plausible path forward, when doing physical experiment­s is very hard.”

Dactyl’s real-world fingers are tracked by infrared dots and cameras. In training, every simulated movement that brought the cube closer to the goal gave Dactyl a small reward. Dropping the cube caused it to feel a penalty 20 times as big.

The process is called reinforcem­ent learning. The robot software repeats the attempts millions of times in a simulated environmen­t, trying over and over to get the highest reward. OpenAI used roughly the same algorithm it used to beat human players in a video game, “Dota 2.”

In real life, a team of researcher­s worked about a year to get the mechanical hand to this point. Why? For one, the hand in a simulated environmen­t doesn’t understand friction. So even though its real fingers are rubbery, Dactyl lacks human understand­ing about the best grips.

Researcher­s injected their simulated environmen­t with changes to gravity, hand angle and other variables so the software learns to operate in a way that is adaptable. That helped narrow the gap between realworld results and simulated ones, which were much better.

The variations helped the hand succeed putting the right letter face up more than a dozen times in a row before dropping the cube. In simulation, the hand typically succeeded 50 times in a row before the test was stopped. (AP)

 ?? (AP) ?? This undated photo provided by OpenAI shows a robotic hand rotating a cube at the company’s research lab in San Francisco.
(AP) This undated photo provided by OpenAI shows a robotic hand rotating a cube at the company’s research lab in San Francisco.

Newspapers in English

Newspapers from Kuwait