Unsupervised Deep Persistent Monocular Visual Odometry and Depth
Estimation in Extreme Environments
- URL: http://arxiv.org/abs/2011.00341v1
- Date: Sat, 31 Oct 2020 19:10:27 GMT
- Title: Unsupervised Deep Persistent Monocular Visual Odometry and Depth
Estimation in Extreme Environments
- Authors: Yasin Almalioglu, Angel Santamaria-Navarro, Benjamin Morrell,
Ali-akbar Agha-mohammadi
- Abstract summary: unsupervised deep learning approaches have received significant attention to estimate the depth and visual odometry (VO) from unlabelled monocular image sequences.
We propose an unsupervised monocular deep VO framework that predicts six-degrees-of-freedom pose camera motion and depth map of the scene from unlabelled RGB image sequences.
The proposed approach outperforms both traditional and state-of-the-art unsupervised deep VO methods providing better results for both pose estimation and depth recovery.
- Score: 7.197188771058501
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, unsupervised deep learning approaches have received
significant attention to estimate the depth and visual odometry (VO) from
unlabelled monocular image sequences. However, their performance is limited in
challenging environments due to perceptual degradation, occlusions and rapid
motions. Moreover, the existing unsupervised methods suffer from the lack of
scale-consistency constraints across frames, which causes that the VO
estimators fail to provide persistent trajectories over long sequences. In this
study, we propose an unsupervised monocular deep VO framework that predicts
six-degrees-of-freedom pose camera motion and depth map of the scene from
unlabelled RGB image sequences. We provide detailed quantitative and
qualitative evaluations of the proposed framework on a) a challenging dataset
collected during the DARPA Subterranean challenge; and b) the benchmark KITTI
and Cityscapes datasets. The proposed approach outperforms both traditional and
state-of-the-art unsupervised deep VO methods providing better results for both
pose estimation and depth recovery. The presented approach is part of the
solution used by the COSTAR team participating at the DARPA Subterranean
Challenge.
Related papers
- The RoboDepth Challenge: Methods and Advancements Towards Robust Depth Estimation [97.63185634482552]
We summarize the winning solutions from the RoboDepth Challenge.
The challenge was designed to facilitate and advance robust OoD depth estimation.
We hope this challenge could lay a solid foundation for future research on robust and reliable depth estimation.
arXiv Detail & Related papers (2023-07-27T17:59:56Z) - BEVScope: Enhancing Self-Supervised Depth Estimation Leveraging
Bird's-Eye-View in Dynamic Scenarios [12.079195812249747]
Current self-supervised depth estimation methods grapple with several limitations.
We present BEVScope, an innovative approach to self-supervised depth estimation.
We propose an adaptive loss function, specifically designed to mitigate the complexities associated with moving objects.
arXiv Detail & Related papers (2023-06-20T15:16:35Z) - MaskingDepth: Masked Consistency Regularization for Semi-supervised
Monocular Depth Estimation [38.09399326203952]
MaskingDepth is a novel semi-supervised learning framework for monocular depth estimation.
It enforces consistency between the strongly-augmented unlabeled data and the pseudo-labels derived from weakly-augmented unlabeled data.
arXiv Detail & Related papers (2022-12-21T06:56:22Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - On Robust Cross-View Consistency in Self-Supervised Monocular Depth Estimation [56.97699793236174]
We study two kinds of robust cross-view consistency in this paper.
We exploit the temporal coherence in both depth feature space and 3D voxel space for self-supervised monocular depth estimation.
Experimental results on several outdoor benchmarks show that our method outperforms current state-of-the-art techniques.
arXiv Detail & Related papers (2022-09-19T03:46:13Z) - Learning Occlusion-Aware Coarse-to-Fine Depth Map for Self-supervised
Monocular Depth Estimation [11.929584800629673]
We propose a novel network to learn an Occlusion-aware Coarse-to-Fine Depth map for self-supervised monocular depth estimation.
The proposed OCFD-Net does not only employ a discrete depth constraint for learning a coarse-level depth map, but also employ a continuous depth constraint for learning a scene depth residual.
arXiv Detail & Related papers (2022-03-21T12:43:42Z) - A high-precision self-supervised monocular visual odometry in foggy
weather based on robust cycled generative adversarial networks and multi-task
learning aided depth estimation [0.0]
This paper proposes a high-precision self-supervised monocular VO, which is specifically designed for navigation in foggy weather.
A cycled generative adversarial network is designed to obtain high-quality self-supervised loss via forcing the forward and backward half-cycle to output consistent estimation.
gradient-based loss and perceptual loss are introduced to eliminate the interference of complex photometric change on self-supervised loss in foggy weather.
arXiv Detail & Related papers (2022-03-09T15:41:57Z) - Unsupervised Monocular Depth Reconstruction of Non-Rigid Scenes [87.91841050957714]
We present an unsupervised monocular framework for dense depth estimation of dynamic scenes.
We derive a training objective that aims to opportunistically preserve pairwise distances between reconstructed 3D points.
Our method provides promising results, demonstrating its capability of reconstructing 3D from challenging videos of non-rigid scenes.
arXiv Detail & Related papers (2020-12-31T16:02:03Z) - DeFeat-Net: General Monocular Depth via Simultaneous Unsupervised
Representation Learning [65.94499390875046]
DeFeat-Net is an approach to simultaneously learn a cross-domain dense feature representation.
Our technique is able to outperform the current state-of-the-art with around 10% reduction in all error measures.
arXiv Detail & Related papers (2020-03-30T13:10:32Z) - Peeking into occluded joints: A novel framework for crowd pose
estimation [88.56203133287865]
OPEC-Net is an Image-Guided Progressive GCN module that estimates invisible joints from an inference perspective.
OCPose is the most complex Occluded Pose dataset with respect to average IoU between adjacent instances.
arXiv Detail & Related papers (2020-03-23T19:32:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.