Sim-and-Real Reinforcement Learning for Manipulation: A Consensus-based
Approach
- URL: http://arxiv.org/abs/2302.13423v2
- Date: Sun, 17 Sep 2023 12:11:02 GMT
- Title: Sim-and-Real Reinforcement Learning for Manipulation: A Consensus-based
Approach
- Authors: Wenxing Liu, Hanlin Niu, Wei Pan, Guido Herrmann, Joaquin Carrasco
- Abstract summary: We propose a Consensus-based Sim-And-Real deep reinforcement learning algorithm (CSAR) for manipulator pick-and-place tasks.
We train the agents in simulators and the real world to get the optimal policies for both sim-and-real worlds.
- Score: 4.684126055213616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sim-and-real training is a promising alternative to sim-to-real training for
robot manipulations. However, the current sim-and-real training is neither
efficient, i.e., slow convergence to the optimal policy, nor effective, i.e.,
sizeable real-world robot data. Given limited time and hardware budgets, the
performance of sim-and-real training is not satisfactory. In this paper, we
propose a Consensus-based Sim-And-Real deep reinforcement learning algorithm
(CSAR) for manipulator pick-and-place tasks, which shows comparable performance
in both sim-and-real worlds. In this algorithm, we train the agents in
simulators and the real world to get the optimal policies for both sim-and-real
worlds. We found two interesting phenomenons: (1) Best policy in simulation is
not the best for sim-and-real training. (2) The more simulation agents, the
better sim-and-real training. The experimental video is available at:
https://youtu.be/mcHJtNIsTEQ.
Related papers
- Dynamics as Prompts: In-Context Learning for Sim-to-Real System Identifications [23.94013806312391]
We propose a novel approach that dynamically adjusts simulation environment parameters online using in-context learning.
We validate our approach across two tasks: object scooping and table air hockey.
Our approach delivers efficient and smooth system identification, advancing the deployment of robots in dynamic real-world scenarios.
arXiv Detail & Related papers (2024-10-27T07:13:38Z) - Overcoming the Sim-to-Real Gap: Leveraging Simulation to Learn to Explore for Real-World RL [25.991354823569033]
We show that in many regimes, while direct sim2real transfer may fail, we can utilize the simulator to learn a set of emphexploratory policies.
In particular, in the setting of low-rank MDPs, we show that coupling these exploratory policies with simple, practical approaches.
This is the first evidence that simulation transfer yields a provable gain in reinforcement learning in settings where direct sim2real transfer fails.
arXiv Detail & Related papers (2024-10-26T19:12:27Z) - DrEureka: Language Model Guided Sim-To-Real Transfer [64.14314476811806]
Transferring policies learned in simulation to the real world is a promising strategy for acquiring robot skills at scale.
In this paper, we investigate using Large Language Models (LLMs) to automate and accelerate sim-to-real design.
Our approach is capable of solving novel robot tasks, such as quadruped balancing and walking atop a yoga ball.
arXiv Detail & Related papers (2024-06-04T04:53:05Z) - TRANSIC: Sim-to-Real Policy Transfer by Learning from Online Correction [25.36756787147331]
Learning in simulation and transferring the learned policy to the real world has the potential to enable generalist robots.
We propose a data-driven approach to enable successful sim-to-real transfer based on a human-in-the-loop framework.
We show that our approach can achieve successful sim-to-real transfer in complex and contact-rich manipulation tasks such as furniture assembly.
arXiv Detail & Related papers (2024-05-16T17:59:07Z) - Provable Sim-to-real Transfer in Continuous Domain with Partial
Observations [39.18274543757048]
Sim-to-real transfer trains RL agents in the simulated environments and then deploys them in the real world.
We show that a popular robust adversarial training algorithm is capable of learning a policy from the simulated environment that is competitive to the optimal policy in the real-world environment.
arXiv Detail & Related papers (2022-10-27T16:37:52Z) - DeXtreme: Transfer of Agile In-hand Manipulation from Simulation to
Reality [64.51295032956118]
We train a policy that can perform robust dexterous manipulation on an anthropomorphic robot hand.
Our work reaffirms the possibilities of sim-to-real transfer for dexterous manipulation in diverse kinds of hardware and simulator setups.
arXiv Detail & Related papers (2022-10-25T01:51:36Z) - Real2Sim or Sim2Real: Robotics Visual Insertion using Deep Reinforcement
Learning and Real2Sim Policy Adaptation [8.992053371569678]
In this work, we solve the insertion task using a pure visual reinforcement learning solution with minimum infrastructure requirement.
We also propose a novel sim2real strategy, Real2Sim, which provides a novel and easier solution in policy adaptation.
arXiv Detail & Related papers (2022-06-06T15:27:25Z) - Robot Learning from Randomized Simulations: A Review [59.992761565399185]
Deep learning has caused a paradigm shift in robotics research, favoring methods that require large amounts of data.
State-of-the-art approaches learn in simulation where data generation is fast as well as inexpensive.
We focus on a technique named 'domain randomization' which is a method for learning from randomized simulations.
arXiv Detail & Related papers (2021-11-01T13:55:41Z) - TrafficSim: Learning to Simulate Realistic Multi-Agent Behaviors [74.67698916175614]
We propose TrafficSim, a multi-agent behavior model for realistic traffic simulation.
In particular, we leverage an implicit latent variable model to parameterize a joint actor policy.
We show TrafficSim generates significantly more realistic and diverse traffic scenarios as compared to a diverse set of baselines.
arXiv Detail & Related papers (2021-01-17T00:29:30Z) - Reactive Long Horizon Task Execution via Visual Skill and Precondition
Models [59.76233967614774]
We describe an approach for sim-to-real training that can accomplish unseen robotic tasks using models learned in simulation to ground components of a simple task planner.
We show an increase in success rate from 91.6% to 98% in simulation and from 10% to 80% success rate in the real-world as compared with naive baselines.
arXiv Detail & Related papers (2020-11-17T15:24:01Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.