Error-Aware Imitation Learning from Teleoperation Data for Mobile
Manipulation
- URL: http://arxiv.org/abs/2112.05251v1
- Date: Thu, 9 Dec 2021 23:54:59 GMT
- Title: Error-Aware Imitation Learning from Teleoperation Data for Mobile
Manipulation
- Authors: Josiah Wong, Albert Tung, Andrey Kurenkov, Ajay Mandlekar, Li Fei-Fei,
Silvio Savarese, Roberto Mart\'in-Mart\'in
- Abstract summary: In mobile manipulation (MM), robots can both navigate within and interact with their environment.
In this work, we explore how to apply imitation learning (IL) to learn continuous visuo-motor policies for MM tasks.
- Score: 54.31414116478024
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In mobile manipulation (MM), robots can both navigate within and interact
with their environment and are thus able to complete many more tasks than
robots only capable of navigation or manipulation. In this work, we explore how
to apply imitation learning (IL) to learn continuous visuo-motor policies for
MM tasks. Much prior work has shown that IL can train visuo-motor policies for
either manipulation or navigation domains, but few works have applied IL to the
MM domain. Doing this is challenging for two reasons: on the data side, current
interfaces make collecting high-quality human demonstrations difficult, and on
the learning side, policies trained on limited data can suffer from covariate
shift when deployed. To address these problems, we first propose Mobile
Manipulation RoboTurk (MoMaRT), a novel teleoperation framework allowing
simultaneous navigation and manipulation of mobile manipulators, and collect a
first-of-its-kind large scale dataset in a realistic simulated kitchen setting.
We then propose a learned error detection system to address the covariate shift
by detecting when an agent is in a potential failure state. We train performant
IL policies and error detectors from this data, and achieve over 45% task
success rate and 85% error detection success rate across multiple multi-stage
tasks when trained on expert data. Codebase, datasets, visualization, and more
available at https://sites.google.com/view/il-for-mm/home.
Related papers
- Zero-Cost Whole-Body Teleoperation for Mobile Manipulation [8.71539730969424]
MoMa-Teleop is a novel teleoperation method that delegates the base motions to a reinforcement learning agent.
We demonstrate that our approach results in a significant reduction in task completion time across a variety of robots and tasks.
arXiv Detail & Related papers (2024-09-23T15:09:45Z) - VITAL: Visual Teleoperation to Enhance Robot Learning through Human-in-the-Loop Corrections [10.49712834719005]
We propose a low-cost visual teleoperation system for bimanual manipulation tasks, called VITAL.
Our approach leverages affordable hardware and visual processing techniques to collect demonstrations.
We enhance the generalizability and robustness of the learned policies by utilizing both real and simulated environments.
arXiv Detail & Related papers (2024-07-30T23:29:47Z) - Offline Imitation Learning Through Graph Search and Retrieval [57.57306578140857]
Imitation learning is a powerful machine learning algorithm for a robot to acquire manipulation skills.
We propose GSR, a simple yet effective algorithm that learns from suboptimal demonstrations through Graph Search and Retrieval.
GSR can achieve a 10% to 30% higher success rate and over 30% higher proficiency compared to baselines.
arXiv Detail & Related papers (2024-07-22T06:12:21Z) - Harmonic Mobile Manipulation [35.82197562695662]
HarmonicMM is an end-to-end learning method that optimize both navigation and manipulation.
Our contributions include a new benchmark for mobile manipulation and the successful deployment with only RGB visual observation.
arXiv Detail & Related papers (2023-12-11T18:54:42Z) - Multi-skill Mobile Manipulation for Object Rearrangement [75.62774690484022]
We study a modular approach to tackle long-horizon mobile manipulation tasks for object rearrangement.
Prior work chains multiple stationary manipulation skills with a point-goal navigation skill, which are learned individually on subtasks.
We operationalize these ideas by implementing mobile manipulation skills rather than stationary ones and training a navigation skill trained with region goal instead of point goal.
arXiv Detail & Related papers (2022-09-06T19:02:08Z) - Human-in-the-Loop Imitation Learning using Remote Teleoperation [72.2847988686463]
We build a data collection system tailored to 6-DoF manipulation settings.
We develop an algorithm to train the policy iteratively on new data collected by the system.
We demonstrate that agents trained on data collected by our intervention-based system and algorithm outperform agents trained on an equivalent number of samples collected by non-interventional demonstrators.
arXiv Detail & Related papers (2020-12-12T05:30:35Z) - COG: Connecting New Skills to Past Experience with Offline Reinforcement
Learning [78.13740204156858]
We show that we can reuse prior data to extend new skills simply through dynamic programming.
We demonstrate the effectiveness of our approach by chaining together several behaviors seen in prior datasets for solving a new task.
We train our policies in an end-to-end fashion, mapping high-dimensional image observations to low-level robot control commands.
arXiv Detail & Related papers (2020-10-27T17:57:29Z) - Visual Imitation Made Easy [102.36509665008732]
We present an alternate interface for imitation that simplifies the data collection process while allowing for easy transfer to robots.
We use commercially available reacher-grabber assistive tools both as a data collection device and as the robot's end-effector.
We experimentally evaluate on two challenging tasks: non-prehensile pushing and prehensile stacking, with 1000 diverse demonstrations for each task.
arXiv Detail & Related papers (2020-08-11T17:58:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.