Unsupervised Visual Odometry and Action Integration for PointGoal
Navigation in Indoor Environment
- URL: http://arxiv.org/abs/2210.00413v2
- Date: Mon, 3 Apr 2023 09:18:10 GMT
- Title: Unsupervised Visual Odometry and Action Integration for PointGoal
Navigation in Indoor Environment
- Authors: Yijun Cao, Xianshi Zhang, Fuya Luo, Chuan Lin, and Yongjie Li
- Abstract summary: PointGoal navigation in indoor environment is a fundamental task for personal robots to navigate to a specified point.
To improve the PointGoal navigation accuracy without GPS signal, we use visual odometry (VO) and propose a novel action integration module (AIM) trained in unsupervised manner.
Experiments show that the proposed system achieves satisfactory results and outperforms the partially supervised learning algorithms on the popular Gibson dataset.
- Score: 14.363948775085534
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: PointGoal navigation in indoor environment is a fundamental task for personal
robots to navigate to a specified point. Recent studies solved this PointGoal
navigation task with near-perfect success rate in photo-realistically simulated
environments, under the assumptions with noiseless actuation and most
importantly, perfect localization with GPS and compass sensors. However,
accurate GPS signalis difficult to be obtained in real indoor environment. To
improve the PointGoal navigation accuracy without GPS signal, we use visual
odometry (VO) and propose a novel action integration module (AIM) trained in
unsupervised manner. Sepecifically, unsupervised VO computes the relative pose
of the agent from the re-projection error of two adjacent frames, and then
replaces the accurate GPS signal with the path integration. The pseudo position
estimated by VO is used to train action integration which assists agent to
update their internal perception of location and helps improve the success rate
of navigation. The training and inference process only use RGB, depth,
collision as well as self-action information. The experiments show that the
proposed system achieves satisfactory results and outperforms the partially
supervised learning algorithms on the popular Gibson dataset.
Related papers
- MPVO: Motion-Prior based Visual Odometry for PointGoal Navigation [3.9974562667271507]
Visual odometry (VO) is essential for enabling accurate point-goal navigation of embodied agents in indoor environments.
Recent deep-learned VO methods show robust performance but suffer from sample inefficiency during training.
We propose a robust and sample-efficient VO pipeline based on motion priors available while an agent is navigating an environment.
arXiv Detail & Related papers (2024-11-07T15:36:49Z) - A Bionic Data-driven Approach for Long-distance Underwater Navigation with Anomaly Resistance [59.21686775951903]
Various animals exhibit accurate navigation using environment cues.
Inspired by animal navigation, this work proposes a bionic and data-driven approach for long-distance underwater navigation.
The proposed approach uses measured geomagnetic data for the navigation, and requires no GPS systems or geographical maps.
arXiv Detail & Related papers (2024-02-06T13:20:56Z) - Angle Robustness Unmanned Aerial Vehicle Navigation in GNSS-Denied
Scenarios [66.05091704671503]
We present a novel angle navigation paradigm to deal with flight deviation in point-to-point navigation tasks.
We also propose a model that includes the Adaptive Feature Enhance Module, Cross-knowledge Attention-guided Module and Robust Task-oriented Head Module.
arXiv Detail & Related papers (2024-02-04T08:41:20Z) - Collaborative Learning with a Drone Orchestrator [79.75113006257872]
A swarm of intelligent wireless devices train a shared neural network model with the help of a drone.
The proposed framework achieves a significant speedup in training, leading to an average 24% and 87% saving in the drone hovering time.
arXiv Detail & Related papers (2023-03-03T23:46:25Z) - Performance Study of YOLOv5 and Faster R-CNN for Autonomous Navigation
around Non-Cooperative Targets [0.0]
This paper discusses how the combination of cameras and machine learning algorithms can achieve the relative navigation task.
The performance of two deep learning-based object detection algorithms, Faster Region-based Convolutional Neural Networks (R-CNN) and You Only Look Once (YOLOv5) is tested.
The paper discusses the path to implementing the feature recognition algorithms and towards integrating them into the spacecraft Guidance Navigation and Control system.
arXiv Detail & Related papers (2023-01-22T04:53:38Z) - ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints [94.60414567852536]
Long-range navigation requires both planning and reasoning about local traversability.
We propose a learning-based approach that integrates learning and planning.
ViKiNG can leverage its image-based learned controller and goal-directed to navigate to goals up to 3 kilometers away.
arXiv Detail & Related papers (2022-02-23T02:14:23Z) - The Surprising Effectiveness of Visual Odometry Techniques for Embodied
PointGoal Navigation [100.08270721713149]
PointGoal navigation has been introduced in simulated Embodied AI environments.
Recent advances solve this PointGoal navigation task with near-perfect accuracy (99.6% success)
We show that integrating visual odometry techniques into navigation policies improves the state-of-the-art on the popular Habitat PointNav benchmark by a large margin.
arXiv Detail & Related papers (2021-08-26T02:12:49Z) - Towards bio-inspired unsupervised representation learning for indoor
aerial navigation [4.26712082692017]
This research displays a biologically inspired deep-learning algorithm for simultaneous localization and mapping (SLAM) and its application in a drone navigation system.
We propose an unsupervised representation learning method that yields low-dimensional latent state descriptors, that mitigates the sensitivity to perceptual aliasing, and works on power-efficient, embedded hardware.
The designed algorithm is evaluated on a dataset collected in an indoor warehouse environment, and initial results show the feasibility for robust indoor aerial navigation.
arXiv Detail & Related papers (2021-06-17T08:42:38Z) - Multimodal Scale Consistency and Awareness for Monocular Self-Supervised
Depth Estimation [1.1470070927586016]
Self-supervised approaches on monocular videos suffer from scale-inconsistency across long sequences.
We propose a dynamically-weighted GPS-to-Scale (g2s) loss to complement the appearance-based losses.
We demonstrate scale-consistent and -aware depth estimation during inference, improving the performance even when training with low-frequency GPS data.
arXiv Detail & Related papers (2021-03-03T15:39:41Z) - Integrating Egocentric Localization for More Realistic Point-Goal
Navigation Agents [90.65480527538723]
We develop point-goal navigation agents that rely on visual estimates of egomotion under noisy action dynamics.
Our agent was the runner-up in the PointNav track of CVPR 2020 Habitat Challenge.
arXiv Detail & Related papers (2020-09-07T16:52:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.