GNM: A General Navigation Model to Drive Any Robot
- URL: http://arxiv.org/abs/2210.03370v2
- Date: Mon, 22 May 2023 04:13:00 GMT
- Title: GNM: A General Navigation Model to Drive Any Robot
- Authors: Dhruv Shah, Ajay Sridhar, Arjun Bhorkar, Noriaki Hirose, Sergey Levine
- Abstract summary: General goal-conditioned model for vision-based navigation can be trained on data obtained from many distinct but structurally similar robots.
We analyze the necessary design decisions for effective data sharing across robots.
We deploy the trained GNM on a range of new robots, including an under quadrotor.
- Score: 67.40225397212717
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning provides a powerful tool for vision-based navigation, but the
capabilities of learning-based policies are constrained by limited training
data. If we could combine data from all available sources, including multiple
kinds of robots, we could train more powerful navigation models. In this paper,
we study how a general goal-conditioned model for vision-based navigation can
be trained on data obtained from many distinct but structurally similar robots,
and enable broad generalization across environments and embodiments. We analyze
the necessary design decisions for effective data sharing across robots,
including the use of temporal context and standardized action spaces, and
demonstrate that an omnipolicy trained from heterogeneous datasets outperforms
policies trained on any single dataset. We curate 60 hours of navigation
trajectories from 6 distinct robots, and deploy the trained GNM on a range of
new robots, including an underactuated quadrotor. We find that training on
diverse data leads to robustness against degradation in sensing and actuation.
Using a pre-trained navigation model with broad generalization capabilities can
bootstrap applications on novel robots going forward, and we hope that the GNM
represents a step in that direction. For more information on the datasets,
code, and videos, please check out our project page
https://sites.google.com/view/drive-any-robot.
Related papers
- Robot Utility Models: General Policies for Zero-Shot Deployment in New Environments [26.66666135624716]
We present Robot Utility Models (RUMs), a framework for training and deploying zero-shot robot policies.
RUMs can generalize to new environments without any finetuning.
We train five utility models for opening cabinet doors, opening drawers, picking up napkins, picking up paper bags, and reorienting fallen objects.
arXiv Detail & Related papers (2024-09-09T17:59:50Z) - Scaling Cross-Embodied Learning: One Policy for Manipulation, Navigation, Locomotion and Aviation [49.03165169369552]
By training a single policy across many different kinds of robots, a robot learning method can leverage much broader and more diverse datasets.
We propose CrossFormer, a scalable and flexible transformer-based policy that can consume data from any embodiment.
We demonstrate that the same network weights can control vastly different robots, including single and dual arm manipulation systems, wheeled robots, quadcopters, and quadrupeds.
arXiv Detail & Related papers (2024-08-21T17:57:51Z) - AutoRT: Embodied Foundation Models for Large Scale Orchestration of Robotic Agents [109.3804962220498]
AutoRT is a system to scale up the deployment of operational robots in completely unseen scenarios with minimal human supervision.
We demonstrate AutoRT proposing instructions to over 20 robots across multiple buildings and collecting 77k real robot episodes via both teleoperation and autonomous robot policies.
We experimentally show that such "in-the-wild" data collected by AutoRT is significantly more diverse, and that AutoRT's use of LLMs allows for instruction following data collection robots that can align to human preferences.
arXiv Detail & Related papers (2024-01-23T18:45:54Z) - ViNT: A Foundation Model for Visual Navigation [52.2571739391896]
Visual Navigation Transformer (ViNT) is a foundation model for vision-based robotic navigation.
ViNT is trained with a general goal-reaching objective that can be used with any navigation dataset.
It exhibits positive transfer, outperforming specialist models trained on singular datasets.
arXiv Detail & Related papers (2023-06-26T16:57:03Z) - RT-1: Robotics Transformer for Real-World Control at Scale [98.09428483862165]
We present a model class, dubbed Robotics Transformer, that exhibits promising scalable model properties.
We verify our conclusions in a study of different model classes and their ability to generalize as a function of the data size, model size, and data diversity based on a large-scale data collection on real robots performing real-world tasks.
arXiv Detail & Related papers (2022-12-13T18:55:15Z) - Multi-Robot Deep Reinforcement Learning for Mobile Navigation [82.62621210336881]
We propose a deep reinforcement learning algorithm with hierarchically integrated models (HInt)
At training time, HInt learns separate perception and dynamics models, and at test time, HInt integrates the two models in a hierarchical manner and plans actions with the integrated model.
Our mobile navigation experiments show that HInt outperforms conventional hierarchical policies and single-source approaches.
arXiv Detail & Related papers (2021-06-24T19:07:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.