Imitate and Repurpose: Learning Reusable Robot Movement Skills From
Human and Animal Behaviors
- URL: http://arxiv.org/abs/2203.17138v1
- Date: Thu, 31 Mar 2022 16:01:32 GMT
- Title: Imitate and Repurpose: Learning Reusable Robot Movement Skills From
Human and Animal Behaviors
- Authors: Steven Bohez, Saran Tunyasuvunakool, Philemon Brakel, Fereshteh
Sadeghi, Leonard Hasenclever, Yuval Tassa, Emilio Parisotto, Jan Humplik,
Tuomas Haarnoja, Roland Hafner, Markus Wulfmeier, Michael Neunert, Ben Moran,
Noah Siegel, Andrea Huber, Francesco Romano, Nathan Batchelor, Federico
Casarini, Josh Merel, Raia Hadsell, Nicolas Heess
- Abstract summary: We investigate the use of prior knowledge of human and animal movement to learn reusable locomotion skills for real legged robots.
Our approach builds upon previous work on imitating human or dog Motion Capture (MoCap) data to learn a movement skill module.
- Score: 28.22210425264389
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We investigate the use of prior knowledge of human and animal movement to
learn reusable locomotion skills for real legged robots. Our approach builds
upon previous work on imitating human or dog Motion Capture (MoCap) data to
learn a movement skill module. Once learned, this skill module can be reused
for complex downstream tasks. Importantly, due to the prior imposed by the
MoCap data, our approach does not require extensive reward engineering to
produce sensible and natural looking behavior at the time of reuse. This makes
it easy to create well-regularized, task-oriented controllers that are suitable
for deployment on real robots. We demonstrate how our skill module can be used
for imitation, and train controllable walking and ball dribbling policies for
both the ANYmal quadruped and OP3 humanoid. These policies are then deployed on
hardware via zero-shot simulation-to-reality transfer. Accompanying videos are
available at https://bit.ly/robot-npmp.
Related papers
- HumanPlus: Humanoid Shadowing and Imitation from Humans [82.47551890765202]
We introduce a full-stack system for humanoids to learn motion and autonomous skills from human data.
We first train a low-level policy in simulation via reinforcement learning using existing 40-hour human motion datasets.
We then perform supervised behavior cloning to train skill policies using egocentric vision, allowing humanoids to complete different tasks autonomously.
arXiv Detail & Related papers (2024-06-15T00:41:34Z) - Expressive Whole-Body Control for Humanoid Robots [20.132927075816742]
We learn a whole-body control policy on a human-sized robot to mimic human motions as realistic as possible.
With training in simulation and Sim2Real transfer, our policy can control a humanoid robot to walk in different styles, shake hands with humans, and even dance with a human in the real world.
arXiv Detail & Related papers (2024-02-26T18:09:24Z) - Universal Humanoid Motion Representations for Physics-Based Control [71.46142106079292]
We present a universal motion representation that encompasses a comprehensive range of motor skills for physics-based humanoid control.
We first learn a motion imitator that can imitate all of human motion from a large, unstructured motion dataset.
We then create our motion representation by distilling skills directly from the imitator.
arXiv Detail & Related papers (2023-10-06T20:48:43Z) - Generalized Animal Imitator: Agile Locomotion with Versatile Motion Prior [14.114972332185044]
This paper introduces the Versatile Motion prior (VIM) - a Reinforcement Learning framework designed to incorporate a range of agile locomotion tasks.
Our framework enables legged robots to learn diverse agile low-level skills by imitating animal motions and manually designed motions.
Our evaluations of the VIM framework span both simulation environments and real-world deployment.
arXiv Detail & Related papers (2023-10-02T17:59:24Z) - Lifelike Agility and Play in Quadrupedal Robots using Reinforcement Learning and Generative Pre-trained Models [28.519964304030236]
We propose a hierarchical framework to construct primitive-, environmental- and strategic-level knowledge that are all pre-trainable, reusable and enrichable for legged robots.
The primitive module summarizes knowledge from animal motion data, where, inspired by large pre-trained models in language and image understanding, we introduce deep generative models to produce motor control signals stimulating legged robots to act like real animals.
We apply the trained hierarchical controllers to the MAX robot, a quadrupedal robot developed in-house, to mimic animals, traverse complex obstacles and play in a designed challenging multi-agent chase tag game.
arXiv Detail & Related papers (2023-08-29T09:22:12Z) - HERD: Continuous Human-to-Robot Evolution for Learning from Human
Demonstration [57.045140028275036]
We show that manipulation skills can be transferred from a human to a robot through the use of micro-evolutionary reinforcement learning.
We propose an algorithm for multi-dimensional evolution path searching that allows joint optimization of both the robot evolution path and the policy.
arXiv Detail & Related papers (2022-12-08T15:56:13Z) - Learning Reward Functions for Robotic Manipulation by Observing Humans [92.30657414416527]
We use unlabeled videos of humans solving a wide range of manipulation tasks to learn a task-agnostic reward function for robotic manipulation policies.
The learned rewards are based on distances to a goal in an embedding space learned using a time-contrastive objective.
arXiv Detail & Related papers (2022-11-16T16:26:48Z) - Adaptation of Quadruped Robot Locomotion with Meta-Learning [64.71260357476602]
We demonstrate that meta-reinforcement learning can be used to successfully train a robot capable to solve a wide range of locomotion tasks.
The performance of the meta-trained robot is similar to that of a robot that is trained on a single task.
arXiv Detail & Related papers (2021-07-08T10:37:18Z) - Learning Bipedal Robot Locomotion from Human Movement [0.791553652441325]
We present a reinforcement learning based method for teaching a real world bipedal robot to perform movements directly from motion capture data.
Our method seamlessly transitions from training in a simulation environment to executing on a physical robot.
We demonstrate our method on an internally developed humanoid robot with movements ranging from a dynamic walk cycle to complex balancing and waving.
arXiv Detail & Related papers (2021-05-26T00:49:37Z) - Learning Agile Robotic Locomotion Skills by Imitating Animals [72.36395376558984]
Reproducing the diverse and agile locomotion skills of animals has been a longstanding challenge in robotics.
We present an imitation learning system that enables legged robots to learn agile locomotion skills by imitating real-world animals.
arXiv Detail & Related papers (2020-04-02T02:56:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.