Learning and Retrieval from Prior Data for Skill-based Imitation
Learning
- URL: http://arxiv.org/abs/2210.11435v1
- Date: Thu, 20 Oct 2022 17:34:59 GMT
- Title: Learning and Retrieval from Prior Data for Skill-based Imitation
Learning
- Authors: Soroush Nasiriany and Tian Gao and Ajay Mandlekar and Yuke Zhu
- Abstract summary: We develop a skill-based imitation learning framework that extracts temporally extended sensorimotor skills from prior data.
We identify several key design choices that significantly improve performance on novel tasks.
- Score: 47.59794569496233
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Imitation learning offers a promising path for robots to learn
general-purpose behaviors, but traditionally has exhibited limited scalability
due to high data supervision requirements and brittle generalization. Inspired
by recent advances in multi-task imitation learning, we investigate the use of
prior data from previous tasks to facilitate learning novel tasks in a robust,
data-efficient manner. To make effective use of the prior data, the robot must
internalize knowledge from past experiences and contextualize this knowledge in
novel tasks. To that end, we develop a skill-based imitation learning framework
that extracts temporally extended sensorimotor skills from prior data and
subsequently learns a policy for the target task that invokes these learned
skills. We identify several key design choices that significantly improve
performance on novel tasks, namely representation learning objectives to enable
more predictable skill representations and a retrieval-based data augmentation
mechanism to increase the scope of supervision for policy training. On a
collection of simulated and real-world manipulation domains, we demonstrate
that our method significantly outperforms existing imitation learning and
offline reinforcement learning approaches. Videos and code are available at
https://ut-austin-rpl.github.io/sailor
Related papers
- Meta-Learning and representation learner: A short theoretical note [0.0]
Meta-learning is a subfield of machine learning where the goal is to develop models and algorithms that can learn from various tasks.
Unlike traditional machine learning methods focusing on learning a specific task, meta-learning aims to leverage experience from previous tasks to enhance future learning.
arXiv Detail & Related papers (2024-07-04T23:47:10Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - Towards A Unified Agent with Foundation Models [18.558328028366816]
We investigate how to embed and leverage such abilities in Reinforcement Learning (RL) agents.
We design a framework that uses language as the core reasoning tool, exploring how this enables an agent to tackle a series of fundamental RL challenges.
We demonstrate substantial performance improvements over baselines in exploration efficiency and ability to reuse data from offline datasets.
arXiv Detail & Related papers (2023-07-18T22:37:30Z) - Skill-based Meta-Reinforcement Learning [65.31995608339962]
We devise a method that enables meta-learning on long-horizon, sparse-reward tasks.
Our core idea is to leverage prior experience extracted from offline datasets during meta-learning.
arXiv Detail & Related papers (2022-04-25T17:58:19Z) - Lifelong Adaptive Machine Learning for Sensor-based Human Activity
Recognition Using Prototypical Networks [0.0]
Continual learning, also known as lifelong learning, is an emerging research topic that has been attracting increasing interest in the field of machine learning.
We build on recent advances in the area of continual machine learning and design a lifelong adaptive learning framework using Prototypical Networks, LAPNet-HAR.
LAPNet-HAR processes sensor-based data streams in a task-free data-incremental fashion and mitigates catastrophic forgetting using experience replay and continual prototype adaptation.
arXiv Detail & Related papers (2022-03-11T00:57:29Z) - Hierarchical Skills for Efficient Exploration [70.62309286348057]
In reinforcement learning, pre-trained low-level skills have the potential to greatly facilitate exploration.
Prior knowledge of the downstream task is required to strike the right balance between generality (fine-grained control) and specificity (faster learning) in skill design.
We propose a hierarchical skill learning framework that acquires skills of varying complexity in an unsupervised manner.
arXiv Detail & Related papers (2021-10-20T22:29:32Z) - Parrot: Data-Driven Behavioral Priors for Reinforcement Learning [79.32403825036792]
We propose a method for pre-training behavioral priors that can capture complex input-output relationships observed in successful trials.
We show how this learned prior can be used for rapidly learning new tasks without impeding the RL agent's ability to try out novel behaviors.
arXiv Detail & Related papers (2020-11-19T18:47:40Z) - COG: Connecting New Skills to Past Experience with Offline Reinforcement
Learning [78.13740204156858]
We show that we can reuse prior data to extend new skills simply through dynamic programming.
We demonstrate the effectiveness of our approach by chaining together several behaviors seen in prior datasets for solving a new task.
We train our policies in an end-to-end fashion, mapping high-dimensional image observations to low-level robot control commands.
arXiv Detail & Related papers (2020-10-27T17:57:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.