RealMonoDepth: Self-Supervised Monocular Depth Estimation for General
Scenes
- URL: http://arxiv.org/abs/2004.06267v1
- Date: Tue, 14 Apr 2020 02:03:10 GMT
- Title: RealMonoDepth: Self-Supervised Monocular Depth Estimation for General
Scenes
- Authors: Mertalp Ocal, Armin Mustafa
- Abstract summary: Existing supervised methods for monocular depth estimation require accurate depth measurements for training.
Self-supervised approaches have demonstrated impressive results but do not generalise to scenes with different depth ranges or camera baselines.
We introduce RealMonoDepth, a self-supervised monocular depth estimation approach which learns to estimate the real scene depth for a diverse range of indoor and outdoor scenes.
- Score: 11.995578248462946
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a generalised self-supervised learning approach for monocular
estimation of the real depth across scenes with diverse depth ranges from
1--100s of meters. Existing supervised methods for monocular depth estimation
require accurate depth measurements for training. This limitation has led to
the introduction of self-supervised methods that are trained on stereo image
pairs with a fixed camera baseline to estimate disparity which is transformed
to depth given known calibration. Self-supervised approaches have demonstrated
impressive results but do not generalise to scenes with different depth ranges
or camera baselines. In this paper, we introduce RealMonoDepth a
self-supervised monocular depth estimation approach which learns to estimate
the real scene depth for a diverse range of indoor and outdoor scenes. A novel
loss function with respect to the true scene depth based on relative depth
scaling and warping is proposed. This allows self-supervised training of a
single network with multiple data sets for scenes with diverse depth ranges
from both stereo pair and in the wild moving camera data sets. A comprehensive
performance evaluation across five benchmark data sets demonstrates that
RealMonoDepth provides a single trained network which generalises depth
estimation across indoor and outdoor scenes, consistently outperforming
previous self-supervised approaches.
Related papers
- ScaleDepth: Decomposing Metric Depth Estimation into Scale Prediction and Relative Depth Estimation [62.600382533322325]
We propose a novel monocular depth estimation method called ScaleDepth.
Our method decomposes metric depth into scene scale and relative depth, and predicts them through a semantic-aware scale prediction module.
Our method achieves metric depth estimation for both indoor and outdoor scenes in a unified framework.
arXiv Detail & Related papers (2024-07-11T05:11:56Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - SelfTune: Metrically Scaled Monocular Depth Estimation through
Self-Supervised Learning [53.78813049373321]
We propose a self-supervised learning method for the pre-trained supervised monocular depth networks to enable metrically scaled depth estimation.
Our approach is useful for various applications such as mobile robot navigation and is applicable to diverse environments.
arXiv Detail & Related papers (2022-03-10T12:28:42Z) - Improving Depth Estimation using Location Information [0.0]
This paper improves the self-supervised deep learning techniques to perform accurate generalized monocular depth estimation.
The main idea is to train the deep model to take into account a sequence of the different frames, each frame is geotagged with its location information.
arXiv Detail & Related papers (2021-12-27T22:30:14Z) - Self-Attention Dense Depth Estimation Network for Unrectified Video
Sequences [6.821598757786515]
LiDAR and radar sensors are the hardware solution for real-time depth estimation.
Deep learning based self-supervised depth estimation methods have shown promising results.
We propose a self-attention based depth and ego-motion network for unrectified images.
arXiv Detail & Related papers (2020-05-28T21:53:53Z) - Single Image Depth Estimation Trained via Depth from Defocus Cues [105.67073923825842]
Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
arXiv Detail & Related papers (2020-01-14T20:22:54Z) - Don't Forget The Past: Recurrent Depth Estimation from Monocular Video [92.84498980104424]
We put three different types of depth estimation into a common framework.
Our method produces a time series of depth maps.
It can be applied to monocular videos only or be combined with different types of sparse depth patterns.
arXiv Detail & Related papers (2020-01-08T16:50:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.