E(2)-Equivariant Graph Planning for Navigation
- URL: http://arxiv.org/abs/2309.13043v2
- Date: Sat, 27 Jan 2024 22:29:30 GMT
- Title: E(2)-Equivariant Graph Planning for Navigation
- Authors: Linfeng Zhao, Hongyu Li, Taskin Padir, Huaizu Jiang, Lawson L.S. Wong
- Abstract summary: We exploit Euclidean symmetry in planning for 2D navigation.
To address the challenges of unstructured environments, we formulate the navigation problem as planning on a geometric graph.
- Score: 26.016209191573605
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning for robot navigation presents a critical and challenging task. The
scarcity and costliness of real-world datasets necessitate efficient learning
approaches. In this letter, we exploit Euclidean symmetry in planning for 2D
navigation, which originates from Euclidean transformations between reference
frames and enables parameter sharing. To address the challenges of unstructured
environments, we formulate the navigation problem as planning on a geometric
graph and develop an equivariant message passing network to perform value
iteration. Furthermore, to handle multi-camera input, we propose a learnable
equivariant layer to lift features to a desired space. We conduct comprehensive
evaluations across five diverse tasks encompassing structured and unstructured
environments, along with maps of known and unknown, given point goals or
semantic goals. Our experiments confirm the substantial benefits on training
efficiency, stability, and generalization. More details can be found at the
project website: https://lhy.xyz/e2-planning/.
Related papers
- Multi-Object Navigation in real environments using hybrid policies [18.52681391843433]
We introduce a hybrid navigation method, which decomposes the problem into two different skills.
We show the advantages of this approach compared to end-to-end methods both in simulation and a real environment.
arXiv Detail & Related papers (2024-01-24T20:41:25Z) - Interactive Semantic Map Representation for Skill-based Visual Object
Navigation [43.71312386938849]
This paper introduces a new representation of a scene semantic map formed during the embodied agent interaction with the indoor environment.
We have implemented this representation into a full-fledged navigation approach called SkillTron.
The proposed approach makes it possible to form both intermediate goals for robot exploration and the final goal for object navigation.
arXiv Detail & Related papers (2023-11-07T16:30:12Z) - How To Not Train Your Dragon: Training-free Embodied Object Goal
Navigation with Semantic Frontiers [94.46825166907831]
We present a training-free solution to tackle the object goal navigation problem in Embodied AI.
Our method builds a structured scene representation based on the classic visual simultaneous localization and mapping (V-SLAM) framework.
Our method propagates semantics on the scene graphs based on language priors and scene statistics to introduce semantic knowledge to the geometric frontiers.
arXiv Detail & Related papers (2023-05-26T13:38:33Z) - ETPNav: Evolving Topological Planning for Vision-Language Navigation in
Continuous Environments [56.194988818341976]
Vision-language navigation is a task that requires an agent to follow instructions to navigate in environments.
We propose ETPNav, which focuses on two critical skills: 1) the capability to abstract environments and generate long-range navigation plans, and 2) the ability of obstacle-avoiding control in continuous environments.
ETPNav yields more than 10% and 20% improvements over prior state-of-the-art on R2R-CE and RxR-CE datasets.
arXiv Detail & Related papers (2023-04-06T13:07:17Z) - Weakly-Supervised Multi-Granularity Map Learning for Vision-and-Language
Navigation [87.52136927091712]
We address a practical yet challenging problem of training robot agents to navigate in an environment following a path described by some language instructions.
To achieve accurate and efficient navigation, it is critical to build a map that accurately represents both spatial location and the semantic information of the environment objects.
We propose a multi-granularity map, which contains both object fine-grained details (e.g., color, texture) and semantic classes, to represent objects more comprehensively.
arXiv Detail & Related papers (2022-10-14T04:23:27Z) - ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints [94.60414567852536]
Long-range navigation requires both planning and reasoning about local traversability.
We propose a learning-based approach that integrates learning and planning.
ViKiNG can leverage its image-based learned controller and goal-directed to navigate to goals up to 3 kilometers away.
arXiv Detail & Related papers (2022-02-23T02:14:23Z) - PONI: Potential Functions for ObjectGoal Navigation with
Interaction-free Learning [125.22462763376993]
We propose Potential functions for ObjectGoal Navigation with Interaction-free learning (PONI)
PONI disentangles the skills of where to look?' for an object and how to navigate to (x, y)?'
arXiv Detail & Related papers (2022-01-25T01:07:32Z) - Towards real-world navigation with deep differentiable planners [0.0]
We train embodied neural networks to plan and navigate unseen complex 3D environments.
We focus on differentiable planners such as Value Iteration Networks (VIN), which are trained offline from safe expert demonstrations.
arXiv Detail & Related papers (2021-08-08T11:29:16Z) - SOON: Scenario Oriented Object Navigation with Graph-based Exploration [102.74649829684617]
The ability to navigate like a human towards a language-guided target from anywhere in a 3D embodied environment is one of the 'holy grail' goals of intelligent robots.
Most visual navigation benchmarks focus on navigating toward a target from a fixed starting point, guided by an elaborate set of instructions that depicts step-by-step.
This approach deviates from real-world problems in which human-only describes what the object and its surrounding look like and asks the robot to start navigation from anywhere.
arXiv Detail & Related papers (2021-03-31T15:01:04Z) - Evolving Graphical Planner: Contextual Global Planning for
Vision-and-Language Navigation [47.79784520827089]
We introduce the Evolving Graphical Planner (EGP), a model that performs global planning for navigation based on raw sensory input.
We evaluate our model on a challenging Vision-and-Language Navigation (VLN) task with photorealistic images and achieve superior performance compared to previous navigation architectures.
arXiv Detail & Related papers (2020-07-11T00:21:05Z) - Deep compositional robotic planners that follow natural language
commands [21.481360281719006]
We show how a sampling-based robotic planner can be augmented to learn to understand a sequence of natural language commands.
Our approach combines a deep network structured according to the parse of a complex command that includes objects, verbs, spatial relations, and attributes.
arXiv Detail & Related papers (2020-02-12T19:56:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.