Multi-Stage Cable Routing through Hierarchical Imitation Learning
- URL: http://arxiv.org/abs/2307.08927v5
- Date: Sat, 13 Jan 2024 07:39:35 GMT
- Title: Multi-Stage Cable Routing through Hierarchical Imitation Learning
- Authors: Jianlan Luo, Charles Xu, Xinyang Geng, Gilbert Feng, Kuan Fang, Liam
Tan, Stefan Schaal, Sergey Levine
- Abstract summary: We study the problem of learning to perform multi-stage robotic manipulation tasks, with applications to cable routing.
We present a system for instantiating this method to learn the cable routing task, and perform evaluations showing great performance.
- Score: 52.66135251744562
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the problem of learning to perform multi-stage robotic manipulation
tasks, with applications to cable routing, where the robot must route a cable
through a series of clips. This setting presents challenges representative of
complex multi-stage robotic manipulation scenarios: handling deformable
objects, closing the loop on visual perception, and handling extended behaviors
consisting of multiple steps that must be executed successfully to complete the
entire task. In such settings, learning individual primitives for each stage
that succeed with a high enough rate to perform a complete temporally extended
task is impractical: if each stage must be completed successfully and has a
non-negligible probability of failure, the likelihood of successful completion
of the entire task becomes negligible. Therefore, successful controllers for
such multi-stage tasks must be able to recover from failure and compensate for
imperfections in low-level controllers by smartly choosing which controllers to
trigger at any given time, retrying, or taking corrective action as needed. To
this end, we describe an imitation learning system that uses vision-based
policies trained from demonstrations at both the lower (motor control) and the
upper (sequencing) level, present a system for instantiating this method to
learn the cable routing task, and perform evaluations showing great performance
in generalizing to very challenging clip placement variations. Supplementary
videos, datasets, and code can be found at
https://sites.google.com/view/cablerouting.
Related papers
- BiKC: Keypose-Conditioned Consistency Policy for Bimanual Robotic Manipulation [48.08416841005715]
We introduce a novel keypose-conditioned consistency policy tailored for bimanual manipulation.
It is a hierarchical imitation learning framework that consists of a high-level keypose predictor and a low-level trajectory generator.
Simulated and real-world experimental results demonstrate that the proposed approach surpasses baseline methods in terms of success rate and operational efficiency.
arXiv Detail & Related papers (2024-06-14T14:49:12Z) - Data-CUBE: Data Curriculum for Instruction-based Sentence Representation
Learning [85.66907881270785]
We propose a data curriculum method, namely Data-CUBE, that arranges the orders of all the multi-task data for training.
In the task level, we aim to find the optimal task order to minimize the total cross-task interference risk.
In the instance level, we measure the difficulty of all instances per task, then divide them into the easy-to-difficult mini-batches for training.
arXiv Detail & Related papers (2024-01-07T18:12:20Z) - Leveraging Sequentiality in Reinforcement Learning from a Single
Demonstration [68.94506047556412]
We propose to leverage a sequential bias to learn control policies for complex robotic tasks using a single demonstration.
We show that DCIL-II can solve with unprecedented sample efficiency some challenging simulated tasks such as humanoid locomotion and stand-up.
arXiv Detail & Related papers (2022-11-09T10:28:40Z) - Wish you were here: Hindsight Goal Selection for long-horizon dexterous
manipulation [14.901636098553848]
Solving tasks with a sparse reward in a sample-efficient manner poses a challenge to modern reinforcement learning.
Existing strategies explore based on task-agnostic goal distributions, which can render the solution of long-horizon tasks impractical.
We extend hindsight relabelling mechanisms to guide exploration along task-specific distributions implied by a small set of successful demonstrations.
arXiv Detail & Related papers (2021-12-01T16:12:32Z) - Multi-Task Learning with Sequence-Conditioned Transporter Networks [67.57293592529517]
We aim to solve multi-task learning through the lens of sequence-conditioning and weighted sampling.
We propose a new suite of benchmark aimed at compositional tasks, MultiRavens, which allows defining custom task combinations.
Second, we propose a vision-based end-to-end system architecture, Sequence-Conditioned Transporter Networks, which augments Goal-Conditioned Transporter Networks with sequence-conditioning and weighted sampling.
arXiv Detail & Related papers (2021-09-15T21:19:11Z) - Reset-Free Reinforcement Learning via Multi-Task Learning: Learning
Dexterous Manipulation Behaviors without Human Intervention [67.1936055742498]
We show that multi-task learning can effectively scale reset-free learning schemes to much more complex problems.
This work shows the ability to learn dexterous manipulation behaviors in the real world with RL without any human intervention.
arXiv Detail & Related papers (2021-04-22T17:38:27Z) - Gradient Surgery for Multi-Task Learning [119.675492088251]
Multi-task learning has emerged as a promising approach for sharing structure across multiple tasks.
The reasons why multi-task learning is so challenging compared to single-task learning are not fully understood.
We propose a form of gradient surgery that projects a task's gradient onto the normal plane of the gradient of any other task that has a conflicting gradient.
arXiv Detail & Related papers (2020-01-19T06:33:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.