Deep Blue vs. Kasparov. Watson vs. Ken Jennings and Brad Rutter. Deepmind vs. Atari. Alpha Go vs. Lee Sedol.
The great machine vs. human competitions over the last few decades left no doubt about who’s the boss.
All of those matches were played in a most courteous fashion. The challenges all involved intellectual pursuits.
But what will happen when AI takes on games of physical contact, when, for instance, robots engage in pushing, shoving or knocking opponents over?
Researchers at DeepMind addressed that issue during trials of humanoid robots trained to play soccer. No human subjects were involved in these contests, not yet anyway. But there was some rough play.
In a paper released last week of the arXiv preprint server, Tuomas Haarnoja and more than two dozen colleagues reported on their successful efforts to teach complex movement skills and basic game strategy to robots.
The researchers said that numerous projects by others in recent years involving quadrupedal robots have yielded impressive results. Notable among them was Boston Dynamics’ robot dog Spot that excelled at smoothly navigating unknown, unstructured and hostile environments.
Fewer projects have tackled bipedal movement. The researchers say two-legged mobility poses additional challenges concerning stability and safety. When it comes to sports, those challenges are even greater.
“Soccer requires a diverse set of highly agile and dynamic movements, including running, turning, side stepping, kicking, passing, fall recovery, object interaction and many more,” Haarnoja said.
“Players further need to be able to make predictions about the ball, teammates and opponents, and adapt their movements to the game context. Players also need to coordinate movements over long time scales to achieve tactical, coordinated play.”
The crew at DeepMind designed miniature humanoid robots with 20 controllable joints and used Deep RL (Deep Reinforcement Learning) to teach them basic soccer skills. They focused on context-adaptive movement skills such as “walking, running, turning, kicking and fall recovery.”
The robots exhibited “robust and dynamic movement skills,” said Haarnoja. The report, titled “Learning Agile Soccer Skills for a Bipedal Robot with Deep Reinforcement Learning,” was also posted on a Google blog last week.
The robotic soccer project differed from many earlier similar projects in that it focused employing the entire robotic body—not just hands or feet—to engage in strategic play.
“Creating general embodied intelligence, that is creating agents that can act in the physical world with agility, dexterity and understanding—as animals or humans do— is one of the long-standing goals of AI researchers and roboticists alike,” Haarnoja said.
Publication of the DeepMind project stirred much discussion on social media, but one brief video clip drew particular attention. In the clip, a researcher is shown continually pushing down a robot trying to score a goal. The robot heroically proceeded each time to recover and get back on its feet.
Although clearly done to test and improve the robot‘s ability to recover from stumbles and other errors, the “abuse” stirred Twitter users to respond.
“It’s hard not to anthropomorphize. My brain says, STOP BEING MEAN! Lol,” said John Weller.
“At what point do [the robots] learn it’s easier to kick the football when a human isn’t constantly pushing them and then take action to stop it happening so they can play ball?” asked N4GERACS.
“I gotta say, personally, given how fast AI is moving these days, I would not be so cavalier about shoving those robots,” Jeff Kirsch warned.
And a cautionary note from Jing Lang: “I hope they will forgive us for this.”
Perhaps those unnerved by the shoving will be comforted by the wisdom of a great athlete from a different sport, the late great baseball slugger Babe Ruth, who once observed, “You can’t beat the person who won’t give up.”
Tuomas Haarnoja et al, Learning Agile Soccer Skills for a Bipedal Robot with Deep Reinforcement Learning, arXiv (2023). DOI: 10.48550/arxiv.2304.13653
Project site: sites.google.com/view/op3-soccer
© 2023 Science X Network
Knocked down but not defeated: Robots learn soccer (2023, May 2)
retrieved 7 May 2023
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.