Towards General and Autonomous Learning of Core Skills: A Case Study in
Locomotion
- URL: http://arxiv.org/abs/2008.12228v1
- Date: Thu, 6 Aug 2020 08:23:55 GMT
- Title: Towards General and Autonomous Learning of Core Skills: A Case Study in
Locomotion
- Authors: Roland Hafner, Tim Hertweck, Philipp Kl\"oppner, Michael Bloesch,
Michael Neunert, Markus Wulfmeier, Saran Tunyasuvunakool, Nicolas Heess,
Martin Riedmiller
- Abstract summary: We develop a learning framework that can learn sophisticated locomotion behavior for a wide spectrum of legged robots.
Our learning framework relies on a data-efficient, off-policy multi-task RL algorithm and a small set of reward functions that are semantically identical across robots.
For nine different types of robots, including a real-world quadruped robot, we demonstrate that the same algorithm can rapidly learn diverse and reusable locomotion skills.
- Score: 19.285099263193622
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern Reinforcement Learning (RL) algorithms promise to solve difficult
motor control problems directly from raw sensory inputs. Their attraction is
due in part to the fact that they can represent a general class of methods that
allow to learn a solution with a reasonably set reward and minimal prior
knowledge, even in situations where it is difficult or expensive for a human
expert. For RL to truly make good on this promise, however, we need algorithms
and learning setups that can work across a broad range of problems with minimal
problem specific adjustments or engineering. In this paper, we study this idea
of generality in the locomotion domain. We develop a learning framework that
can learn sophisticated locomotion behavior for a wide spectrum of legged
robots, such as bipeds, tripeds, quadrupeds and hexapods, including wheeled
variants. Our learning framework relies on a data-efficient, off-policy
multi-task RL algorithm and a small set of reward functions that are
semantically identical across robots. To underline the general applicability of
the method, we keep the hyper-parameter settings and reward definitions
constant across experiments and rely exclusively on on-board sensing. For nine
different types of robots, including a real-world quadruped robot, we
demonstrate that the same algorithm can rapidly learn diverse and reusable
locomotion skills without any platform specific adjustments or additional
instrumentation of the learning setup.
Related papers
- Multi-Objective Algorithms for Learning Open-Ended Robotic Problems [1.0124625066746598]
Quadrupedal locomotion is a complex, open-ended problem vital to expanding autonomous vehicle reach.
Traditional reinforcement learning approaches often fall short due to training instability and sample inefficiency.
We propose a novel method leveraging multi-objective evolutionary algorithms as an automatic curriculum learning mechanism.
arXiv Detail & Related papers (2024-11-11T16:26:42Z) - One Policy to Run Them All: an End-to-end Learning Approach to Multi-Embodiment Locomotion [18.556470359899855]
We introduce URMA, the Unified Robot Morphology Architecture.
Our framework brings the end-to-end Multi-Task Reinforcement Learning approach to the realm of legged robots.
We show that URMA can learn a locomotion policy on multiple embodiments that can be easily transferred to unseen robot platforms.
arXiv Detail & Related papers (2024-09-10T09:44:15Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Decentralized Motor Skill Learning for Complex Robotic Systems [5.669790037378093]
We propose a Decentralized motor skill (DEMOS) learning algorithm to automatically discover motor groups that can be decoupled from each other.
Our method improves the robustness and generalization of the policy without sacrificing performance.
Experiments on quadruped and humanoid robots demonstrate that the learned policy is robust against local motor malfunctions and can be transferred to new tasks.
arXiv Detail & Related papers (2023-06-30T05:55:34Z) - Learning and Adapting Agile Locomotion Skills by Transferring Experience [71.8926510772552]
We propose a framework for training complex robotic skills by transferring experience from existing controllers to jumpstart learning new tasks.
We show that our method enables learning complex agile jumping behaviors, navigating to goal locations while walking on hind legs, and adapting to new environments.
arXiv Detail & Related papers (2023-04-19T17:37:54Z) - A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free
Reinforcement Learning [86.06110576808824]
Deep reinforcement learning is a promising approach to learning policies in uncontrolled environments.
Recent advancements in machine learning algorithms and libraries combined with a carefully tuned robot controller lead to learning quadruped in only 20 minutes in the real world.
arXiv Detail & Related papers (2022-08-16T17:37:36Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Learning Agile Robotic Locomotion Skills by Imitating Animals [72.36395376558984]
Reproducing the diverse and agile locomotion skills of animals has been a longstanding challenge in robotics.
We present an imitation learning system that enables legged robots to learn agile locomotion skills by imitating real-world animals.
arXiv Detail & Related papers (2020-04-02T02:56:16Z) - Learning to Walk in the Real World with Minimal Human Effort [80.7342153519654]
We develop a system for learning legged locomotion policies with deep RL in the real world with minimal human effort.
Our system can automatically and efficiently learn locomotion skills on a Minitaur robot with little human intervention.
arXiv Detail & Related papers (2020-02-20T03:36:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.