REvolveR: Continuous Evolutionary Models for Robot-to-robot Policy
Transfer
- URL: http://arxiv.org/abs/2202.05244v1
- Date: Thu, 10 Feb 2022 18:50:25 GMT
- Title: REvolveR: Continuous Evolutionary Models for Robot-to-robot Policy
Transfer
- Authors: Xingyu Liu, Deepak Pathak, Kris M. Kitani
- Abstract summary: We consider the problem of transferring a policy across two different robots with significantly different parameters such as kinematics and morphology.
Existing approaches that train a new policy by matching the action or state transition distribution, including imitation learning methods, fail due to optimal action and/or state distribution being mismatched in different robots.
We propose a novel method named $REvolveR$ of using continuous evolutionary models for robotic policy transfer implemented in a physics simulator.
- Score: 57.045140028275036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A popular paradigm in robotic learning is to train a policy from scratch for
every new robot. This is not only inefficient but also often impractical for
complex robots. In this work, we consider the problem of transferring a policy
across two different robots with significantly different parameters such as
kinematics and morphology. Existing approaches that train a new policy by
matching the action or state transition distribution, including imitation
learning methods, fail due to optimal action and/or state distribution being
mismatched in different robots. In this paper, we propose a novel method named
$REvolveR$ of using continuous evolutionary models for robotic policy transfer
implemented in a physics simulator. We interpolate between the source robot and
the target robot by finding a continuous evolutionary change of robot
parameters. An expert policy on the source robot is transferred through
training on a sequence of intermediate robots that gradually evolve into the
target robot. Experiments show that the proposed continuous evolutionary model
can effectively transfer the policy across robots and achieve superior sample
efficiency on new robots using a physics simulator. The proposed method is
especially advantageous in sparse reward settings where exploration can be
significantly reduced.
Related papers
- Meta-Evolve: Continuous Robot Evolution for One-to-many Policy Transfer [68.10957584496866]
We propose a method that uses continuous robot evolution to efficiently transfer the policy to each target robot.
The robot evolution tree allows the robot evolution paths to be shared, so our approach can significantly outperform naive one-to-one policy transfer.
arXiv Detail & Related papers (2024-05-06T14:52:23Z) - RoboGen: Towards Unleashing Infinite Data for Automated Robot Learning via Generative Simulation [68.70755196744533]
RoboGen is a generative robotic agent that automatically learns diverse robotic skills at scale via generative simulation.
Our work attempts to extract the extensive and versatile knowledge embedded in large-scale models and transfer them to the field of robotics.
arXiv Detail & Related papers (2023-11-02T17:59:21Z) - Human-Robot Skill Transfer with Enhanced Compliance via Dynamic Movement
Primitives [1.7901837062462316]
We introduce a systematic method to extract the dynamic features from human demonstration to auto-tune the parameters in the Dynamic Movement Primitives framework.
Our method was implemented into an actual human-robot setup to extract human dynamic features and used to regenerate the robot trajectories following both LfD and RL.
arXiv Detail & Related papers (2023-04-12T08:48:28Z) - Learning Bipedal Walking for Humanoids with Current Feedback [5.429166905724048]
We present an approach for overcoming the sim2real gap issue for humanoid robots arising from inaccurate torque-tracking at the actuator level.
Our approach successfully trains a unified, end-to-end policy in simulation that can be deployed on a real HRP-5P humanoid robot to achieve bipedal locomotion.
arXiv Detail & Related papers (2023-03-07T08:16:46Z) - HERD: Continuous Human-to-Robot Evolution for Learning from Human
Demonstration [57.045140028275036]
We show that manipulation skills can be transferred from a human to a robot through the use of micro-evolutionary reinforcement learning.
We propose an algorithm for multi-dimensional evolution path searching that allows joint optimization of both the robot evolution path and the policy.
arXiv Detail & Related papers (2022-12-08T15:56:13Z) - GenLoco: Generalized Locomotion Controllers for Quadrupedal Robots [87.32145104894754]
We introduce a framework for training generalized locomotion (GenLoco) controllers for quadrupedal robots.
Our framework synthesizes general-purpose locomotion controllers that can be deployed on a large variety of quadrupedal robots.
We show that our models acquire more general control strategies that can be directly transferred to novel simulated and real-world robots.
arXiv Detail & Related papers (2022-09-12T15:14:32Z) - Learning Bipedal Robot Locomotion from Human Movement [0.791553652441325]
We present a reinforcement learning based method for teaching a real world bipedal robot to perform movements directly from motion capture data.
Our method seamlessly transitions from training in a simulation environment to executing on a physical robot.
We demonstrate our method on an internally developed humanoid robot with movements ranging from a dynamic walk cycle to complex balancing and waving.
arXiv Detail & Related papers (2021-05-26T00:49:37Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - Learning Locomotion Skills in Evolvable Robots [10.167123492952694]
We introduce a controller architecture and a generic learning method to allow a modular robot with an arbitrary shape to learn to walk towards a target and follow this target if it moves.
Our approach is validated on three robots, a spider, a gecko, and their offspring, in three real-world scenarios.
arXiv Detail & Related papers (2020-10-19T14:01:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.