SGoLAM: Simultaneous Goal Localization and Mapping for Multi-Object Goal
Navigation
- URL: http://arxiv.org/abs/2110.07171v1
- Date: Thu, 14 Oct 2021 06:15:14 GMT
- Title: SGoLAM: Simultaneous Goal Localization and Mapping for Multi-Object Goal
Navigation
- Authors: Junho Kim, Eun Sun Lee, Mingi Lee, Donsu Zhang, and Young Min Kim
- Abstract summary: We present SGoLAM, a simple and efficient algorithm for Multi-Object Goal navigation.
Given an agent equipped with an RGB-D camera and a GPS/ sensor, our objective is to have the agent navigate to a sequence of target objects in realistic 3D environments.
SGoLAM is ranked 2nd in the CVPR 2021 MultiON (Multi-Object Goal Navigation) challenge.
- Score: 5.447924312563365
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present SGoLAM, short for simultaneous goal localization and mapping,
which is a simple and efficient algorithm for Multi-Object Goal navigation.
Given an agent equipped with an RGB-D camera and a GPS/Compass sensor, our
objective is to have the agent navigate to a sequence of target objects in
realistic 3D environments. Our pipeline fully leverages the strength of
classical approaches for visual navigation, by decomposing the problem into two
key components: mapping and goal localization. The mapping module converts the
depth observations into an occupancy map, and the goal localization module
marks the locations of goal objects. The agent's policy is determined using the
information provided by the two modules: if a current goal is found, plan
towards the goal and otherwise, perform exploration. As our approach does not
require any training of neural networks, it could be used in an off-the-shelf
manner, and amenable for fast generalization in new, unseen environments.
Nonetheless, our approach performs on par with the state-of-the-art
learning-based approaches. SGoLAM is ranked 2nd in the CVPR 2021 MultiON
(Multi-Object Goal Navigation) challenge. We have made our code publicly
available at \emph{https://github.com/eunsunlee/SGoLAM}.
Related papers
- Diffusion as Reasoning: Enhancing Object Goal Navigation with LLM-Biased Diffusion Model [9.939998139837426]
We propose a new approach to solving the ObjectNav task, by training a diffusion model to learn the statistical distribution patterns of objects in semantic maps.
We also propose the global target bias and local LLM bias methods, where the former can constrain the diffusion model to generate the target object more effectively.
Based on the generated map in the unknown region, the agent sets the predicted location of the target as the goal and moves towards it.
arXiv Detail & Related papers (2024-10-29T08:10:06Z) - IPPON: Common Sense Guided Informative Path Planning for Object Goal Navigation [33.979481250363584]
This paper introduces a novel informative path planning and 3D object probability mapping approach.
The mapping module computes the probability of the object of interest through semantic segmentation and a Bayes filter.
Although our planner follows a zero-shot approach, it achieves state-of-the-art performance as measured by the Success weighted by Path Length (SPL) and Soft SPL in the Habitat ObjectNav Challenge 2023.
arXiv Detail & Related papers (2024-10-25T17:11:33Z) - GOMAA-Geo: GOal Modality Agnostic Active Geo-localization [49.599465495973654]
We consider the task of active geo-localization (AGL) in which an agent uses a sequence of visual cues observed during aerial navigation to find a target specified through multiple possible modalities.
GOMAA-Geo is a goal modality active geo-localization agent for zero-shot generalization between different goal modalities.
arXiv Detail & Related papers (2024-06-04T02:59:36Z) - Zero-Shot Object Goal Visual Navigation With Class-Independent Relationship Network [3.0820097046465285]
"Zero-shot" means that the target the agent needs to find is not trained during the training phase.
We propose the Class-Independent Relationship Network (CIRN) to address the issue of coupling navigation ability with target features during training.
Our method outperforms the current state-of-the-art approaches in the zero-shot object goal visual navigation task.
arXiv Detail & Related papers (2023-10-15T16:42:14Z) - NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration [57.15811390835294]
This paper describes how we can train a single unified diffusion policy to handle both goal-directed navigation and goal-agnostic exploration.
We show that this unified policy results in better overall performance when navigating to visually indicated goals in novel environments.
Our experiments, conducted on a real-world mobile robot platform, show effective navigation in unseen environments in comparison with five alternative methods.
arXiv Detail & Related papers (2023-10-11T21:07:14Z) - Unleash the Potential of Image Branch for Cross-modal 3D Object
Detection [67.94357336206136]
We present a new cross-modal 3D object detector, namely UPIDet, which aims to unleash the potential of the image branch from two aspects.
First, UPIDet introduces a new 2D auxiliary task called normalized local coordinate map estimation.
Second, we discover that the representational capability of the point cloud backbone can be enhanced through the gradients backpropagated from the training objectives of the image branch.
arXiv Detail & Related papers (2023-01-22T08:26:58Z) - SOON: Scenario Oriented Object Navigation with Graph-based Exploration [102.74649829684617]
The ability to navigate like a human towards a language-guided target from anywhere in a 3D embodied environment is one of the 'holy grail' goals of intelligent robots.
Most visual navigation benchmarks focus on navigating toward a target from a fixed starting point, guided by an elaborate set of instructions that depicts step-by-step.
This approach deviates from real-world problems in which human-only describes what the object and its surrounding look like and asks the robot to start navigation from anywhere.
arXiv Detail & Related papers (2021-03-31T15:01:04Z) - Object Goal Navigation using Goal-Oriented Semantic Exploration [98.14078233526476]
This work studies the problem of object goal navigation which involves navigating to an instance of the given object category in unseen environments.
We propose a modular system called, Goal-Oriented Semantic Exploration' which builds an episodic semantic map and uses it to explore the environment efficiently.
arXiv Detail & Related papers (2020-07-01T17:52:32Z) - Improving Target-driven Visual Navigation with Attention on 3D Spatial
Relationships [52.72020203771489]
We investigate target-driven visual navigation using deep reinforcement learning (DRL) in 3D indoor scenes.
Our proposed method combines visual features and 3D spatial representations to learn navigation policy.
Our experiments, performed in the AI2-THOR, show that our model outperforms the baselines in both SR and SPL metrics.
arXiv Detail & Related papers (2020-04-29T08:46:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.