On Embodied Visual Navigation in Real Environments Through Habitat
- URL: http://arxiv.org/abs/2010.13439v1
- Date: Mon, 26 Oct 2020 09:19:07 GMT
- Title: On Embodied Visual Navigation in Real Environments Through Habitat
- Authors: Marco Rosano, Antonino Furnari, Luigi Gulino, Giovanni Maria Farinella
- Abstract summary: Visual navigation models based on deep learning can learn effective policies when trained on large amounts of visual observations.
To deal with this limitation, several simulation platforms have been proposed in order to train visual navigation policies on virtual environments efficiently.
We show that our tool can effectively help to train and evaluate navigation policies on real-world observations without running navigation pisodes in the real world.
- Score: 20.630139085937586
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual navigation models based on deep learning can learn effective policies
when trained on large amounts of visual observations through reinforcement
learning. Unfortunately, collecting the required experience in the real world
requires the deployment of a robotic platform, which is expensive and
time-consuming. To deal with this limitation, several simulation platforms have
been proposed in order to train visual navigation policies on virtual
environments efficiently. Despite the advantages they offer, simulators present
a limited realism in terms of appearance and physical dynamics, leading to
navigation policies that do not generalize in the real world.
In this paper, we propose a tool based on the Habitat simulator which
exploits real world images of the environment, together with sensor and
actuator noise models, to produce more realistic navigation episodes. We
perform a range of experiments to assess the ability of such policies to
generalize using virtual and real-world images, as well as observations
transformed with unsupervised domain adaptation approaches. We also assess the
impact of sensor and actuation noise on the navigation performance and
investigate whether it allows to learn more robust navigation policies. We show
that our tool can effectively help to train and evaluate navigation policies on
real-world observations without running navigation pisodes in the real world.
Related papers
- Learning autonomous driving from aerial imagery [67.06858775696453]
Photogrammetric simulators allow the synthesis of novel views through the transformation of pre-generated assets into novel views.
We use a Neural Radiance Field (NeRF) as an intermediate representation to synthesize novel views from the point of view of a ground vehicle.
arXiv Detail & Related papers (2024-10-18T05:09:07Z) - Learning Navigational Visual Representations with Semantic Map
Supervision [85.91625020847358]
We propose a navigational-specific visual representation learning method by contrasting the agent's egocentric views and semantic maps.
Ego$2$-Map learning transfers the compact and rich information from a map, such as objects, structure and transition, to the agent's egocentric representations for navigation.
arXiv Detail & Related papers (2023-07-23T14:01:05Z) - Virtual Guidance as a Mid-level Representation for Navigation [8.712750753534532]
"Virtual Guidance" is designed to visually represent non-visual instructional signals.
We evaluate our proposed method through experiments in both simulated and real-world settings.
arXiv Detail & Related papers (2023-03-05T17:55:15Z) - Navigating to Objects in the Real World [76.1517654037993]
We present a large-scale empirical study of semantic visual navigation methods comparing methods from classical, modular, and end-to-end learning approaches.
We find that modular learning works well in the real world, attaining a 90% success rate.
In contrast, end-to-end learning does not, dropping from 77% simulation to 23% real-world success rate due to a large image domain gap between simulation and reality.
arXiv Detail & Related papers (2022-12-02T01:10:47Z) - Image-based Navigation in Real-World Environments via Multiple Mid-level
Representations: Fusion Models, Benchmark and Efficient Evaluation [13.207579081178716]
In recent learning-based navigation approaches, the scene understanding and navigation abilities of the agent are achieved simultaneously.
Unfortunately, even if simulators represent an efficient tool to train navigation policies, the resulting models often fail when transferred into the real world.
One possible solution is to provide the navigation model with mid-level visual representations containing important domain-invariant properties of the scene.
arXiv Detail & Related papers (2022-02-02T15:00:44Z) - ViNG: Learning Open-World Navigation with Visual Goals [82.84193221280216]
We propose a learning-based navigation system for reaching visually indicated goals.
We show that our system, which we call ViNG, outperforms previously-proposed methods for goal-conditioned reinforcement learning.
We demonstrate ViNG on a number of real-world applications, such as last-mile delivery and warehouse inspection.
arXiv Detail & Related papers (2020-12-17T18:22:32Z) - Unsupervised Domain Adaptation for Visual Navigation [115.85181329193092]
We propose an unsupervised domain adaptation method for visual navigation.
Our method translates the images in the target domain to the source domain such that the translation is consistent with the representations learned by the navigation policy.
arXiv Detail & Related papers (2020-10-27T18:22:43Z) - Embodied Visual Navigation with Automatic Curriculum Learning in Real
Environments [20.017277077448924]
NavACL is a method of automatic curriculum learning tailored to the navigation task.
Deep reinforcement learning agents trained using NavACL significantly outperform state-of-the-art agents trained with uniform sampling.
Our agents can navigate through unknown cluttered indoor environments to semantically-specified targets using only RGB images.
arXiv Detail & Related papers (2020-09-11T13:28:26Z) - Visual Navigation Among Humans with Optimal Control as a Supervisor [72.5188978268463]
We propose an approach that combines learning-based perception with model-based optimal control to navigate among humans.
Our approach is enabled by our novel data-generation tool, HumANav.
We demonstrate that the learned navigation policies can anticipate and react to humans without explicitly predicting future human motion.
arXiv Detail & Related papers (2020-03-20T16:13:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.