Reset-Free Reinforcement Learning via Multi-Task Learning: Learning
Dexterous Manipulation Behaviors without Human Intervention
- URL: http://arxiv.org/abs/2104.11203v1
- Date: Thu, 22 Apr 2021 17:38:27 GMT
- Title: Reset-Free Reinforcement Learning via Multi-Task Learning: Learning
Dexterous Manipulation Behaviors without Human Intervention
- Authors: Abhishek Gupta, Justin Yu, Tony Z. Zhao, Vikash Kumar, Aaron Rovinsky,
Kelvin Xu, Thomas Devlin, Sergey Levine
- Abstract summary: We show that multi-task learning can effectively scale reset-free learning schemes to much more complex problems.
This work shows the ability to learn dexterous manipulation behaviors in the real world with RL without any human intervention.
- Score: 67.1936055742498
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement Learning (RL) algorithms can in principle acquire complex
robotic skills by learning from large amounts of data in the real world,
collected via trial and error. However, most RL algorithms use a carefully
engineered setup in order to collect data, requiring human supervision and
intervention to provide episodic resets. This is particularly evident in
challenging robotics problems, such as dexterous manipulation. To make data
collection scalable, such applications require reset-free algorithms that are
able to learn autonomously, without explicit instrumentation or human
intervention. Most prior work in this area handles single-task learning.
However, we might also want robots that can perform large repertoires of
skills. At first, this would appear to only make the problem harder. However,
the key observation we make in this work is that an appropriately chosen
multi-task RL setting actually alleviates the reset-free learning challenge,
with minimal additional machinery required. In effect, solving a multi-task
problem can directly solve the reset-free problem since different combinations
of tasks can serve to perform resets for other tasks. By learning multiple
tasks together and appropriately sequencing them, we can effectively learn all
of the tasks together reset-free. This type of multi-task learning can
effectively scale reset-free learning schemes to much more complex problems, as
we demonstrate in our experiments. We propose a simple scheme for multi-task
learning that tackles the reset-free learning problem, and show its
effectiveness at learning to solve complex dexterous manipulation tasks in both
hardware and simulation without any explicit resets. This work shows the
ability to learn dexterous manipulation behaviors in the real world with RL
without any human intervention.
Related papers
- REBOOT: Reuse Data for Bootstrapping Efficient Real-World Dexterous
Manipulation [61.7171775202833]
We introduce an efficient system for learning dexterous manipulation skills withReinforcement learning.
The main idea of our approach is the integration of recent advances in sample-efficient RL and replay buffer bootstrapping.
Our system completes the real-world training cycle by incorporating learned resets via an imitation-based pickup policy.
arXiv Detail & Related papers (2023-09-06T19:05:31Z) - Hindsight States: Blending Sim and Real Task Elements for Efficient
Reinforcement Learning [61.3506230781327]
In robotics, one approach to generate training data builds on simulations based on dynamics models derived from first principles.
Here, we leverage the imbalance in complexity of the dynamics to learn more sample-efficiently.
We validate our method on several challenging simulated tasks and demonstrate that it improves learning both alone and when combined with an existing hindsight algorithm.
arXiv Detail & Related papers (2023-03-03T21:55:04Z) - Don't Start From Scratch: Leveraging Prior Data to Automate Robotic
Reinforcement Learning [70.70104870417784]
Reinforcement learning (RL) algorithms hold the promise of enabling autonomous skill acquisition for robotic systems.
In practice, real-world robotic RL typically requires time consuming data collection and frequent human intervention to reset the environment.
In this work, we study how these challenges can be tackled by effective utilization of diverse offline datasets collected from previously seen tasks.
arXiv Detail & Related papers (2022-07-11T08:31:22Z) - Lifelong Robotic Reinforcement Learning by Retaining Experiences [61.79346922421323]
Many multi-task reinforcement learning efforts assume the robot can collect data from all tasks at all times.
In this work, we study a practical sequential multi-task RL problem motivated by the practical constraints of physical robotic systems.
We derive an approach that effectively leverages the data and policies learned for previous tasks to cumulatively grow the robot's skill-set.
arXiv Detail & Related papers (2021-09-19T18:00:51Z) - Continual Learning of Control Primitives: Skill Discovery via
Reset-Games [128.36174682118488]
We show how a single method can allow an agent to acquire skills with minimal supervision.
We do this by exploiting the insight that the need to "reset" an agent to a broad set of initial states for a learning task provides a natural setting to learn a diverse set of "reset-skills"
arXiv Detail & Related papers (2020-11-10T18:07:44Z) - Curriculum Learning with Hindsight Experience Replay for Sequential
Object Manipulation Tasks [1.370633147306388]
We present an algorithm that combines curriculum learning with Hindsight Experience Replay (HER) to learn sequential object manipulation tasks.
The algorithm exploits the recurrent structure inherent in many object manipulation tasks and implements the entire learning process in the original simulation without adjusting it to each source task.
arXiv Detail & Related papers (2020-08-21T08:59:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.