Hierarchical Robot Navigation in Novel Environments using Rough 2-D Maps
- URL: http://arxiv.org/abs/2106.03665v1
- Date: Mon, 7 Jun 2021 14:42:51 GMT
- Title: Hierarchical Robot Navigation in Novel Environments using Rough 2-D Maps
- Authors: Chengguang Xu, Christopher Amato, Lawson L.S. Wong
- Abstract summary: We propose an approach that leverages a rough 2-D map of the environment to navigate in novel environments without requiring further learning.
Because the low-level controller is only trained with local behaviors, this framework allows us to generalize to novel environments.
Experimental results demonstrate the effectiveness of the proposed framework in both seen and novel environments.
- Score: 21.245942227850733
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In robot navigation, generalizing quickly to unseen environments is
essential. Hierarchical methods inspired by human navigation have been
proposed, typically consisting of a high-level landmark proposer and a
low-level controller. However, these methods either require precise high-level
information to be given in advance or need to construct such guidance from
extensive interaction with the environment. In this work, we propose an
approach that leverages a rough 2-D map of the environment to navigate in novel
environments without requiring further learning. In particular, we introduce a
dynamic topological map that can be initialized from the rough 2-D map along
with a high-level planning approach for proposing reachable 2-D map patches of
the intermediate landmarks between the start and goal locations. To use
proposed 2-D patches, we train a deep generative model to generate intermediate
landmarks in observation space which are used as subgoals by low-level
goal-conditioned reinforcement learning. Importantly, because the low-level
controller is only trained with local behaviors (e.g. go across the
intersection, turn left at a corner) on existing environments, this framework
allows us to generalize to novel environments given only a rough 2-D map,
without requiring further learning. Experimental results demonstrate the
effectiveness of the proposed framework in both seen and novel environments.
Related papers
- Mapping High-level Semantic Regions in Indoor Environments without
Object Recognition [50.624970503498226]
The present work proposes a method for semantic region mapping via embodied navigation in indoor environments.
To enable region identification, the method uses a vision-to-language model to provide scene information for mapping.
By projecting egocentric scene understanding into the global frame, the proposed method generates a semantic map as a distribution over possible region labels at each location.
arXiv Detail & Related papers (2024-03-11T18:09:50Z) - TopoNav: Topological Navigation for Efficient Exploration in Sparse Reward Environments [0.6597195879147555]
TopoNav is a novel framework for efficient goal-oriented exploration and navigation in sparse-reward settings.
TopoNav dynamically constructs a topological map of the environment, capturing key locations and pathways.
We evaluate TopoNav both in the simulated and real-world off-road environments using a Clearpath Jackal robot.
arXiv Detail & Related papers (2024-02-06T15:05:25Z) - NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration [57.15811390835294]
This paper describes how we can train a single unified diffusion policy to handle both goal-directed navigation and goal-agnostic exploration.
We show that this unified policy results in better overall performance when navigating to visually indicated goals in novel environments.
Our experiments, conducted on a real-world mobile robot platform, show effective navigation in unseen environments in comparison with five alternative methods.
arXiv Detail & Related papers (2023-10-11T21:07:14Z) - ETPNav: Evolving Topological Planning for Vision-Language Navigation in
Continuous Environments [56.194988818341976]
Vision-language navigation is a task that requires an agent to follow instructions to navigate in environments.
We propose ETPNav, which focuses on two critical skills: 1) the capability to abstract environments and generate long-range navigation plans, and 2) the ability of obstacle-avoiding control in continuous environments.
ETPNav yields more than 10% and 20% improvements over prior state-of-the-art on R2R-CE and RxR-CE datasets.
arXiv Detail & Related papers (2023-04-06T13:07:17Z) - Think Global, Act Local: Dual-scale Graph Transformer for
Vision-and-Language Navigation [87.03299519917019]
We propose a dual-scale graph transformer (DUET) for joint long-term action planning and fine-grained cross-modal understanding.
We build a topological map on-the-fly to enable efficient exploration in global action space.
The proposed approach, DUET, significantly outperforms state-of-the-art methods on goal-oriented vision-and-language navigation benchmarks.
arXiv Detail & Related papers (2022-02-23T19:06:53Z) - Landmark-Guided Subgoal Generation in Hierarchical Reinforcement
Learning [64.97599673479678]
We present HIerarchical reinforcement learning Guided by Landmarks (HIGL)
HIGL is a novel framework for training a high-level policy with a reduced action space guided by landmarks.
Our experiments demonstrate that our framework outperforms prior-arts across a variety of control tasks.
arXiv Detail & Related papers (2021-10-26T12:16:19Z) - Deep Reinforcement Learning for Adaptive Exploration of Unknown
Environments [6.90777229452271]
We develop an adaptive exploration approach to trade off between exploration and exploitation in one single step for UAVs.
The proposed approach uses a map segmentation technique to decompose the environment map into smaller, tractable maps.
The results demonstrate that our proposed approach is capable of navigating through randomly generated environments and covering more AoI in less time steps compared to the baselines.
arXiv Detail & Related papers (2021-05-04T16:29:44Z) - SOON: Scenario Oriented Object Navigation with Graph-based Exploration [102.74649829684617]
The ability to navigate like a human towards a language-guided target from anywhere in a 3D embodied environment is one of the 'holy grail' goals of intelligent robots.
Most visual navigation benchmarks focus on navigating toward a target from a fixed starting point, guided by an elaborate set of instructions that depicts step-by-step.
This approach deviates from real-world problems in which human-only describes what the object and its surrounding look like and asks the robot to start navigation from anywhere.
arXiv Detail & Related papers (2021-03-31T15:01:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.