Fleet-DAgger: Interactive Robot Fleet Learning with Scalable Human
Supervision
- URL: http://arxiv.org/abs/2206.14349v1
- Date: Wed, 29 Jun 2022 01:23:57 GMT
- Title: Fleet-DAgger: Interactive Robot Fleet Learning with Scalable Human
Supervision
- Authors: Ryan Hoque, Lawrence Yunliang Chen, Satvik Sharma, Karthik
Dharmarajan, Brijen Thananjeyan, Pieter Abbeel, Ken Goldberg
- Abstract summary: Commercial and industrial deployments of robot fleets often fall back on remote human teleoperators during execution.
We formalize the Interactive Fleet Learning (IFL) setting, in which multiple robots interactively query and learn from multiple human supervisors.
We propose Fleet-DAgger, a family of IFL algorithms, and compare a novel Fleet-DAgger algorithm to 4 baselines in simulation.
- Score: 72.4735163268491
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Commercial and industrial deployments of robot fleets often fall back on
remote human teleoperators during execution when robots are at risk or unable
to make task progress. With continual learning, interventions from the remote
pool of humans can also be used to improve the robot fleet control policy over
time. A central question is how to effectively allocate limited human attention
to individual robots. Prior work addresses this in the single-robot,
single-human setting. We formalize the Interactive Fleet Learning (IFL)
setting, in which multiple robots interactively query and learn from multiple
human supervisors. We present a fully implemented open-source IFL benchmark
suite of GPU-accelerated Isaac Gym environments for the evaluation of IFL
algorithms. We propose Fleet-DAgger, a family of IFL algorithms, and compare a
novel Fleet-DAgger algorithm to 4 baselines in simulation. We also perform 1000
trials of a physical block-pushing experiment with 4 ABB YuMi robot arms.
Experiments suggest that the allocation of humans to robots significantly
affects robot fleet performance, and that our algorithm achieves up to 8.8x
higher return on human effort than baselines. See
https://tinyurl.com/fleet-dagger for code, videos, and supplemental material.
Related papers
- OpenBot-Fleet: A System for Collective Learning with Real Robots [45.739144410591805]
We introduce OpenBot-Fleet, a comprehensive open-source cloud robotics system for navigation.
OpenBot-Fleet uses smartphones for sensing, local compute and communication, Google for secure cloud storage and off-board compute.
In experiments we distribute 72 robots to a crowd of workers who operate them in homes, and show that OpenBot-Fleet can learn robust navigation policies.
arXiv Detail & Related papers (2024-05-13T07:22:50Z) - HumanoidBench: Simulated Humanoid Benchmark for Whole-Body Locomotion and Manipulation [50.616995671367704]
We present a high-dimensional, simulated robot learning benchmark, HumanoidBench, featuring a humanoid robot equipped with dexterous hands.
Our findings reveal that state-of-the-art reinforcement learning algorithms struggle with most tasks, whereas a hierarchical learning approach achieves superior performance when supported by robust low-level policies.
arXiv Detail & Related papers (2024-03-15T17:45:44Z) - Interactive Robot Learning from Verbal Correction [42.37176329867376]
OLAF allows users to teach a robot using verbal corrections when the robot makes mistakes.
A key feature of OLAF is its ability to update the robot's visuomotor neural policy based on the verbal feedback.
We demonstrate the efficacy of our design in experiments where a user teaches a robot to perform long-horizon manipulation tasks.
arXiv Detail & Related papers (2023-10-26T16:46:12Z) - ImitationNet: Unsupervised Human-to-Robot Motion Retargeting via Shared Latent Space [9.806227900768926]
This paper introduces a novel deep-learning approach for human-to-robot motion.
Our method does not require paired human-to-robot data, which facilitates its translation to new robots.
Our model outperforms existing works regarding human-to-robot similarity in terms of efficiency and precision.
arXiv Detail & Related papers (2023-09-11T08:55:04Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Decentralised construction of a global coordinate system in a large
swarm of minimalistic robots [0.8701566919381223]
In this study, we present an algorithm to enable positional self-awareness in a swarm of minimalistic error-prone robots.
Despite being unable to measure the bearing of incoming messages, the robots running our algorithm can calculate their position within a swarm deployed in a regular formation.
Our solution has fewer requirements than state-of-the-art algorithms and contains collective noise-filtering mechanisms.
arXiv Detail & Related papers (2023-02-28T14:14:17Z) - GenLoco: Generalized Locomotion Controllers for Quadrupedal Robots [87.32145104894754]
We introduce a framework for training generalized locomotion (GenLoco) controllers for quadrupedal robots.
Our framework synthesizes general-purpose locomotion controllers that can be deployed on a large variety of quadrupedal robots.
We show that our models acquire more general control strategies that can be directly transferred to novel simulated and real-world robots.
arXiv Detail & Related papers (2022-09-12T15:14:32Z) - Model Predictive Control for Fluid Human-to-Robot Handovers [50.72520769938633]
Planning motions that take human comfort into account is not a part of the human-robot handover process.
We propose to generate smooth motions via an efficient model-predictive control framework.
We conduct human-to-robot handover experiments on a diverse set of objects with several users.
arXiv Detail & Related papers (2022-03-31T23:08:20Z) - Know Thyself: Transferable Visuomotor Control Through Robot-Awareness [22.405839096833937]
Training visuomotor robot controllers from scratch on a new robot typically requires generating large amounts of robot-specific data.
We propose a "robot-aware" solution paradigm that exploits readily available robot "self-knowledge"
Our experiments on tabletop manipulation tasks in simulation and on real robots demonstrate that these plug-in improvements dramatically boost the transferability of visuomotor controllers.
arXiv Detail & Related papers (2021-07-19T17:56:04Z) - Show Me What You Can Do: Capability Calibration on Reachable Workspace
for Human-Robot Collaboration [83.4081612443128]
We show that a short calibration using REMP can effectively bridge the gap between what a non-expert user thinks a robot can reach and the ground-truth.
We show that this calibration procedure not only results in better user perception, but also promotes more efficient human-robot collaborations.
arXiv Detail & Related papers (2021-03-06T09:14:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.