Calibrating Self-supervised Monocular Depth Estimation
- URL: http://arxiv.org/abs/2009.07714v2
- Date: Wed, 13 Oct 2021 12:41:43 GMT
- Title: Calibrating Self-supervised Monocular Depth Estimation
- Authors: Robert McCraith, Lukas Neumann, Andrea Vedaldi
- Abstract summary: In the recent years, many methods demonstrated the ability of neural networks to learn depth and pose changes in a sequence of images, using only self-supervision as the training signal.
We show that incorporating prior information about the camera configuration and the environment, we can remove the scale ambiguity and predict depth directly, still using the self-supervised formulation and not relying on any additional sensors.
- Score: 77.77696851397539
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the recent years, many methods demonstrated the ability of neural networks
to learn depth and pose changes in a sequence of images, using only
self-supervision as the training signal. Whilst the networks achieve good
performance, the often over-looked detail is that due to the inherent ambiguity
of monocular vision they predict depth up to an unknown scaling factor. The
scaling factor is then typically obtained from the LiDAR ground truth at test
time, which severely limits practical applications of these methods. In this
paper, we show that incorporating prior information about the camera
configuration and the environment, we can remove the scale ambiguity and
predict depth directly, still using the self-supervised formulation and not
relying on any additional sensors.
Related papers
- Uncertainty and Self-Supervision in Single-View Depth [0.8158530638728501]
Single-view depth estimation is an ill-posed problem because there are multiple solutions that explain 3D geometry from a single view.
Deep neural networks have been shown to be effective at capturing depth from a single view, but the majority of current methodologies are deterministic in nature.
We have addressed this problem by quantifying the uncertainty of supervised single-view depth for Bayesian deep neural networks.
arXiv Detail & Related papers (2024-06-20T11:46:17Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - Image Masking for Robust Self-Supervised Monocular Depth Estimation [12.435468563991174]
Self-supervised monocular depth estimation is a salient task for 3D scene understanding.
We propose MIMDepth, a method that adapts masked image modeling for self-supervised monocular depth estimation.
arXiv Detail & Related papers (2022-10-05T15:57:53Z) - SelfTune: Metrically Scaled Monocular Depth Estimation through
Self-Supervised Learning [53.78813049373321]
We propose a self-supervised learning method for the pre-trained supervised monocular depth networks to enable metrically scaled depth estimation.
Our approach is useful for various applications such as mobile robot navigation and is applicable to diverse environments.
arXiv Detail & Related papers (2022-03-10T12:28:42Z) - Probabilistic and Geometric Depth: Detecting Objects in Perspective [78.00922683083776]
3D object detection is an important capability needed in various practical applications such as driver assistance systems.
Monocular 3D detection, as an economical solution compared to conventional settings relying on binocular vision or LiDAR, has drawn increasing attention recently but still yields unsatisfactory results.
This paper first presents a systematic study on this problem and observes that the current monocular 3D detection problem can be simplified as an instance depth estimation problem.
arXiv Detail & Related papers (2021-07-29T16:30:33Z) - The Temporal Opportunist: Self-Supervised Multi-Frame Monocular Depth [28.06671063873351]
ManyDepth is an adaptive approach to dense depth estimation.
We present a novel consistency loss that encourages the network to ignore the cost volume when it is deemed unreliable.
arXiv Detail & Related papers (2021-04-29T17:53:42Z) - Variational Monocular Depth Estimation for Reliability Prediction [12.951621755732544]
Self-supervised learning for monocular depth estimation is widely investigated as an alternative to supervised learning approach.
Previous works have successfully improved the accuracy of depth estimation by modifying the model structure.
In this paper, we theoretically formulate a variational model for the monocular depth estimation to predict the reliability of the estimated depth image.
arXiv Detail & Related papers (2020-11-24T06:23:51Z) - Unsupervised Monocular Depth Learning with Integrated Intrinsics and
Spatio-Temporal Constraints [61.46323213702369]
This work presents an unsupervised learning framework that is able to predict at-scale depth maps and egomotion.
Our results demonstrate strong performance when compared to the current state-of-the-art on multiple sequences of the KITTI driving dataset.
arXiv Detail & Related papers (2020-11-02T22:26:58Z) - Learning Monocular Dense Depth from Events [53.078665310545745]
Event cameras produce brightness changes in the form of a stream of asynchronous events instead of intensity frames.
Recent learning-based approaches have been applied to event-based data, such as monocular depth prediction.
We propose a recurrent architecture to solve this task and show significant improvement over standard feed-forward methods.
arXiv Detail & Related papers (2020-10-16T12:36:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.