Expanding Versatility of Agile Locomotion through Policy Transitions
Using Latent State Representation
- URL: http://arxiv.org/abs/2306.08224v1
- Date: Wed, 14 Jun 2023 03:30:04 GMT
- Title: Expanding Versatility of Agile Locomotion through Policy Transitions
Using Latent State Representation
- Authors: Guilherme Christmann, Ying-Sheng Luo, Jonathan Hans Soeseno, Wei-Chao
Chen
- Abstract summary: We propose a robust transition strategy that expands the versatility of robot locomotion in the real-world setting.
Our approach is effective in the real-world and achieves a 19% higher average success rate for the most challenging transition pairs in our experiments.
- Score: 3.8271803328378677
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper proposes the transition-net, a robust transition strategy that
expands the versatility of robot locomotion in the real-world setting. To this
end, we start by distributing the complexity of different gaits into dedicated
locomotion policies applicable to real-world robots. Next, we expand the
versatility of the robot by unifying the policies with robust transitions into
a single coherent meta-controller by examining the latent state
representations. Our approach enables the robot to iteratively expand its skill
repertoire and robustly transition between any policy pair in a library. In our
framework, adding new skills does not introduce any process that alters the
previously learned skills. Moreover, training of a locomotion policy takes less
than an hour with a single consumer GPU. Our approach is effective in the
real-world and achieves a 19% higher average success rate for the most
challenging transition pairs in our experiments compared to existing
approaches.
Related papers
- One Policy to Run Them All: an End-to-end Learning Approach to Multi-Embodiment Locomotion [18.556470359899855]
We introduce URMA, the Unified Robot Morphology Architecture.
Our framework brings the end-to-end Multi-Task Reinforcement Learning approach to the realm of legged robots.
We show that URMA can learn a locomotion policy on multiple embodiments that can be easily transferred to unseen robot platforms.
arXiv Detail & Related papers (2024-09-10T09:44:15Z) - Meta-Evolve: Continuous Robot Evolution for One-to-many Policy Transfer [68.10957584496866]
We propose a method that uses continuous robot evolution to efficiently transfer the policy to each target robot.
The robot evolution tree allows the robot evolution paths to be shared, so our approach can significantly outperform naive one-to-one policy transfer.
arXiv Detail & Related papers (2024-05-06T14:52:23Z) - Distilling Reinforcement Learning Policies for Interpretable Robot Locomotion: Gradient Boosting Machines and Symbolic Regression [53.33734159983431]
This paper introduces a novel approach to distill neural RL policies into more interpretable forms.
We train expert neural network policies using RL and distill them into (i) GBMs, (ii) EBMs, and (iii) symbolic policies.
arXiv Detail & Related papers (2024-03-21T11:54:45Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Advanced Skills by Learning Locomotion and Local Navigation End-to-End [10.872193480485596]
In this work, we propose to solve the complete problem by training an end-to-end policy with deep reinforcement learning.
We demonstrate the successful deployment of policies on a real quadrupedal robot.
arXiv Detail & Related papers (2022-09-26T16:35:00Z) - REvolveR: Continuous Evolutionary Models for Robot-to-robot Policy
Transfer [57.045140028275036]
We consider the problem of transferring a policy across two different robots with significantly different parameters such as kinematics and morphology.
Existing approaches that train a new policy by matching the action or state transition distribution, including imitation learning methods, fail due to optimal action and/or state distribution being mismatched in different robots.
We propose a novel method named $REvolveR$ of using continuous evolutionary models for robotic policy transfer implemented in a physics simulator.
arXiv Detail & Related papers (2022-02-10T18:50:25Z) - An Adaptable Approach to Learn Realistic Legged Locomotion without
Examples [38.81854337592694]
This work proposes a generic approach for ensuring realism in locomotion by guiding the learning process with the spring-loaded inverted pendulum model as a reference.
We present experimental results showing that even in a model-free setup, the learned policies can generate realistic and energy-efficient locomotion gaits for a bipedal and a quadrupedal robot.
arXiv Detail & Related papers (2021-10-28T10:14:47Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - Rapidly Adaptable Legged Robots via Evolutionary Meta-Learning [65.88200578485316]
We present a new meta-learning method that allows robots to quickly adapt to changes in dynamics.
Our method significantly improves adaptation to changes in dynamics in high noise settings.
We validate our approach on a quadruped robot that learns to walk while subject to changes in dynamics.
arXiv Detail & Related papers (2020-03-02T22:56:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.