Bottom-Up Skill Discovery from Unsegmented Demonstrations for
Long-Horizon Robot Manipulation
- URL: http://arxiv.org/abs/2109.13841v1
- Date: Tue, 28 Sep 2021 16:18:54 GMT
- Title: Bottom-Up Skill Discovery from Unsegmented Demonstrations for
Long-Horizon Robot Manipulation
- Authors: Yifeng Zhu, Peter Stone, Yuke Zhu
- Abstract summary: We tackle real-world long-horizon robot manipulation tasks through skill discovery.
We present a bottom-up approach to learning a library of reusable skills from unsegmented demonstrations.
Our method has shown superior performance over state-of-the-art imitation learning methods in multi-stage manipulation tasks.
- Score: 55.31301153979621
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We tackle real-world long-horizon robot manipulation tasks through skill
discovery. We present a bottom-up approach to learning a library of reusable
skills from unsegmented demonstrations and use these skills to synthesize
prolonged robot behaviors. Our method starts with constructing a hierarchical
task structure from each demonstration through agglomerative clustering. From
the task structures of multi-task demonstrations, we identify skills based on
the recurring patterns and train goal-conditioned sensorimotor policies with
hierarchical imitation learning. Finally, we train a meta controller to compose
these skills to solve long-horizon manipulation tasks. The entire model can be
trained on a small set of human demonstrations collected within 30 minutes
without further annotations, making it amendable to real-world deployment. We
systematically evaluated our method in simulation environments and on a real
robot. Our method has shown superior performance over state-of-the-art
imitation learning methods in multi-stage manipulation tasks. Furthermore,
skills discovered from multi-task demonstrations boost the average task success
by $8\%$ compared to those discovered from individual tasks.
Related papers
- SPIRE: Synergistic Planning, Imitation, and Reinforcement Learning for Long-Horizon Manipulation [58.14969377419633]
We propose spire, a system that decomposes tasks into smaller learning subproblems and second combines imitation and reinforcement learning to maximize their strengths.
We find that spire outperforms prior approaches that integrate imitation learning, reinforcement learning, and planning by 35% to 50% in average task performance.
arXiv Detail & Related papers (2024-10-23T17:42:07Z) - Unsupervised Skill Discovery for Robotic Manipulation through Automatic Task Generation [17.222197596599685]
We propose a Skill Learning approach that discovers composable behaviors by solving a large number of autonomously generated tasks.
Our method learns skills allowing the robot to consistently and robustly interact with objects in its environment.
The learned skills can be used to solve a set of unseen manipulation tasks, in simulation as well as on a real robotic platform.
arXiv Detail & Related papers (2024-10-07T09:19:13Z) - Augmented Reality Demonstrations for Scalable Robot Imitation Learning [25.026589453708347]
This paper presents an innovative solution: an Augmented Reality (AR)-assisted framework for demonstration collection.
We empower non-roboticist users to produce demonstrations for robot IL using devices like the HoloLens 2.
We validate our approach with experiments on three classical robotics tasks: reach, push, and pick-and-place.
arXiv Detail & Related papers (2024-03-20T18:30:12Z) - Learning Reward Functions for Robotic Manipulation by Observing Humans [92.30657414416527]
We use unlabeled videos of humans solving a wide range of manipulation tasks to learn a task-agnostic reward function for robotic manipulation policies.
The learned rewards are based on distances to a goal in an embedding space learned using a time-contrastive objective.
arXiv Detail & Related papers (2022-11-16T16:26:48Z) - Continual Learning from Demonstration of Robotics Skills [5.573543601558405]
Methods for teaching motion skills to robots focus on training for a single skill at a time.
We propose an approach for continual learning from demonstration using hypernetworks and neural ordinary differential equation solvers.
arXiv Detail & Related papers (2022-02-14T16:26:52Z) - Learning Multi-Stage Tasks with One Demonstration via Self-Replay [9.34061107793983]
We introduce a novel method to learn everyday-like multi-stage tasks from a single human demonstration.
Inspired by the recent Coarse-to-Fine Imitation Learning method, we model imitation learning as a learned object reaching phase followed by an open-loop replay of the demonstrator's actions.
We evaluate with real-world experiments on a set of everyday-like multi-stage tasks, which we show that our method can solve from a single demonstration.
arXiv Detail & Related papers (2021-11-14T20:57:52Z) - Lifelong Robotic Reinforcement Learning by Retaining Experiences [61.79346922421323]
Many multi-task reinforcement learning efforts assume the robot can collect data from all tasks at all times.
In this work, we study a practical sequential multi-task RL problem motivated by the practical constraints of physical robotic systems.
We derive an approach that effectively leverages the data and policies learned for previous tasks to cumulatively grow the robot's skill-set.
arXiv Detail & Related papers (2021-09-19T18:00:51Z) - MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale [103.7609761511652]
We show how a large-scale collective robotic learning system can acquire a repertoire of behaviors simultaneously.
New tasks can be continuously instantiated from previously learned tasks.
We train and evaluate our system on a set of 12 real-world tasks with data collected from 7 robots.
arXiv Detail & Related papers (2021-04-16T16:38:02Z) - Visual Imitation Made Easy [102.36509665008732]
We present an alternate interface for imitation that simplifies the data collection process while allowing for easy transfer to robots.
We use commercially available reacher-grabber assistive tools both as a data collection device and as the robot's end-effector.
We experimentally evaluate on two challenging tasks: non-prehensile pushing and prehensile stacking, with 1000 diverse demonstrations for each task.
arXiv Detail & Related papers (2020-08-11T17:58:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.