Don't Start From Scratch: Leveraging Prior Data to Automate Robotic
Reinforcement Learning
- URL: http://arxiv.org/abs/2207.04703v1
- Date: Mon, 11 Jul 2022 08:31:22 GMT
- Title: Don't Start From Scratch: Leveraging Prior Data to Automate Robotic
Reinforcement Learning
- Authors: Homer Walke, Jonathan Yang, Albert Yu, Aviral Kumar, Jedrzej Orbik,
Avi Singh, Sergey Levine
- Abstract summary: Reinforcement learning (RL) algorithms hold the promise of enabling autonomous skill acquisition for robotic systems.
In practice, real-world robotic RL typically requires time consuming data collection and frequent human intervention to reset the environment.
In this work, we study how these challenges can be tackled by effective utilization of diverse offline datasets collected from previously seen tasks.
- Score: 70.70104870417784
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning (RL) algorithms hold the promise of enabling
autonomous skill acquisition for robotic systems. However, in practice,
real-world robotic RL typically requires time consuming data collection and
frequent human intervention to reset the environment. Moreover, robotic
policies learned with RL often fail when deployed beyond the carefully
controlled setting in which they were learned. In this work, we study how these
challenges can all be tackled by effective utilization of diverse offline
datasets collected from previously seen tasks. When faced with a new task, our
system adapts previously learned skills to quickly learn to both perform the
new task and return the environment to an initial state, effectively performing
its own environment reset. Our empirical results demonstrate that incorporating
prior data into robotic reinforcement learning enables autonomous learning,
substantially improves sample-efficiency of learning, and enables better
generalization.
Related papers
- So You Think You Can Scale Up Autonomous Robot Data Collection? [22.7035324720716]
Reinforcement learning (RL) comes with the promise of enabling autonomous data collection.
It remains challenging to scale in the real-world partly due to the significant effort required for environment design and instrumentation.
Imitation learning (IL) methods require little to no environment design effort, but instead require significant human supervision.
arXiv Detail & Related papers (2024-11-04T05:31:35Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - REBOOT: Reuse Data for Bootstrapping Efficient Real-World Dexterous
Manipulation [61.7171775202833]
We introduce an efficient system for learning dexterous manipulation skills withReinforcement learning.
The main idea of our approach is the integration of recent advances in sample-efficient RL and replay buffer bootstrapping.
Our system completes the real-world training cycle by incorporating learned resets via an imitation-based pickup policy.
arXiv Detail & Related papers (2023-09-06T19:05:31Z) - Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from
Offline Data [101.43350024175157]
Self-supervised learning has the potential to decrease the amount of human annotation and engineering effort required to learn control strategies.
Our work builds on prior work showing that the reinforcement learning (RL) itself can be cast as a self-supervised problem.
We demonstrate that a self-supervised RL algorithm based on contrastive learning can solve real-world, image-based robotic manipulation tasks.
arXiv Detail & Related papers (2023-06-06T01:36:56Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Scalable Multi-Task Imitation Learning with Autonomous Improvement [159.9406205002599]
We build an imitation learning system that can continuously improve through autonomous data collection.
We leverage the robot's own trials as demonstrations for tasks other than the one that the robot actually attempted.
In contrast to prior imitation learning approaches, our method can autonomously collect data with sparse supervision for continuous improvement.
arXiv Detail & Related papers (2020-02-25T18:56:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.