DiPE: Deeper into Photometric Errors for Unsupervised Learning of Depth
and Ego-motion from Monocular Videos
- URL: http://arxiv.org/abs/2003.01360v3
- Date: Fri, 20 Nov 2020 06:31:22 GMT
- Title: DiPE: Deeper into Photometric Errors for Unsupervised Learning of Depth
and Ego-motion from Monocular Videos
- Authors: Hualie Jiang, Laiyan Ding, Zhenglong Sun and Rui Huang
- Abstract summary: This paper shows that carefully manipulating photometric errors can tackle these difficulties better.
The primary improvement is achieved by a statistical technique that can mask out the invisible or nonstationary pixels in the photometric error map.
We also propose an efficient weighted multi-scale scheme to reduce the artifacts in the predicted depth maps.
- Score: 9.255509741319583
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised learning of depth and ego-motion from unlabelled monocular
videos has recently drawn great attention, which avoids the use of expensive
ground truth in the supervised one. It achieves this by using the photometric
errors between the target view and the synthesized views from its adjacent
source views as the loss. Despite significant progress, the learning still
suffers from occlusion and scene dynamics. This paper shows that carefully
manipulating photometric errors can tackle these difficulties better. The
primary improvement is achieved by a statistical technique that can mask out
the invisible or nonstationary pixels in the photometric error map and thus
prevents misleading the networks. With this outlier masking approach, the depth
of objects moving in the opposite direction to the camera can be estimated more
accurately. To the best of our knowledge, such scenarios have not been
seriously considered in the previous works, even though they pose a higher risk
in applications like autonomous driving. We also propose an efficient weighted
multi-scale scheme to reduce the artifacts in the predicted depth maps.
Extensive experiments on the KITTI dataset show the effectiveness of the
proposed approaches. The overall system achieves state-of-theart performance on
both depth and ego-motion estimation.
Related papers
- Uncertainty and Self-Supervision in Single-View Depth [0.8158530638728501]
Single-view depth estimation is an ill-posed problem because there are multiple solutions that explain 3D geometry from a single view.
Deep neural networks have been shown to be effective at capturing depth from a single view, but the majority of current methodologies are deterministic in nature.
We have addressed this problem by quantifying the uncertainty of supervised single-view depth for Bayesian deep neural networks.
arXiv Detail & Related papers (2024-06-20T11:46:17Z) - Towards Scale-Aware, Robust, and Generalizable Unsupervised Monocular
Depth Estimation by Integrating IMU Motion Dynamics [74.1720528573331]
Unsupervised monocular depth and ego-motion estimation has drawn extensive research attention in recent years.
We propose DynaDepth, a novel scale-aware framework that integrates information from vision and IMU motion dynamics.
We validate the effectiveness of DynaDepth by conducting extensive experiments and simulations on the KITTI and Make3D datasets.
arXiv Detail & Related papers (2022-07-11T07:50:22Z) - Unsupervised Monocular Depth Perception: Focusing on Moving Objects [5.489557739480878]
In this paper, we show that deliberately manipulating photometric errors can efficiently deal with difficulties better.
We first propose an outlier masking technique that considers the occluded or dynamic pixels as statistical outliers in the photometric error map.
With the outlier masking, the network learns the depth of objects that move in the opposite direction to the camera more accurately.
arXiv Detail & Related papers (2021-08-30T08:45:02Z) - Self-supervised Learning of Occlusion Aware Flow Guided 3D Geometry
Perception with Adaptive Cross Weighted Loss from Monocular Videos [5.481942307939029]
Self-supervised deep learning-based 3D scene understanding methods can overcome the difficulty of acquiring the densely labeled ground-truth.
In this paper, we explore the learnable occlusion aware optical flow guided self-supervised depth and camera pose estimation.
Our method shows promising results on KITTI, Make3D, and Cityscapes datasets under multiple tasks.
arXiv Detail & Related papers (2021-08-09T09:21:24Z) - Unsupervised Learning of Monocular Depth and Ego-Motion Using Multiple
Masks [14.82498499423046]
A new unsupervised learning method of depth and ego-motion using multiple masks from monocular video is proposed in this paper.
The depth estimation network and the ego-motion estimation network are trained according to the constraints of depth and ego-motion without truth values.
The experiments on KITTI dataset show our method achieves good performance in terms of depth and ego-motion.
arXiv Detail & Related papers (2021-04-01T12:29:23Z) - Unsupervised Depth and Ego-motion Estimation for Monocular Thermal Video
using Multi-spectral Consistency Loss [76.77673212431152]
We propose an unsupervised learning method for the all-day depth and ego-motion estimation.
The proposed method exploits multi-spectral consistency loss to gives complementary supervision for the networks.
Networks trained with the proposed method robustly estimate the depth and pose from monocular thermal video under low-light and even zero-light conditions.
arXiv Detail & Related papers (2021-03-01T05:29:04Z) - Progressive Depth Learning for Single Image Dehazing [56.71963910162241]
Existing dehazing methods often ignore the depth cues and fail in distant areas where heavier haze disturbs the visibility.
We propose a deep end-to-end model that iteratively estimates image depths and transmission maps.
Our approach benefits from explicitly modeling the inner relationship of image depth and transmission map, which is especially effective for distant hazy areas.
arXiv Detail & Related papers (2021-02-21T05:24:18Z) - SAFENet: Self-Supervised Monocular Depth Estimation with Semantic-Aware
Feature Extraction [27.750031877854717]
We propose SAFENet that is designed to leverage semantic information to overcome the limitations of the photometric loss.
Our key idea is to exploit semantic-aware depth features that integrate the semantic and geometric knowledge.
Experiments on KITTI dataset demonstrate that our methods compete or even outperform the state-of-the-art methods.
arXiv Detail & Related papers (2020-10-06T17:22:25Z) - Adaptive confidence thresholding for monocular depth estimation [83.06265443599521]
We propose a new approach to leverage pseudo ground truth depth maps of stereo images generated from self-supervised stereo matching methods.
The confidence map of the pseudo ground truth depth map is estimated to mitigate performance degeneration by inaccurate pseudo depth maps.
Experimental results demonstrate superior performance to state-of-the-art monocular depth estimation methods.
arXiv Detail & Related papers (2020-09-27T13:26:16Z) - Calibrating Self-supervised Monocular Depth Estimation [77.77696851397539]
In the recent years, many methods demonstrated the ability of neural networks to learn depth and pose changes in a sequence of images, using only self-supervision as the training signal.
We show that incorporating prior information about the camera configuration and the environment, we can remove the scale ambiguity and predict depth directly, still using the self-supervised formulation and not relying on any additional sensors.
arXiv Detail & Related papers (2020-09-16T14:35:45Z) - Neural Ray Surfaces for Self-Supervised Learning of Depth and Ego-motion [51.19260542887099]
We show that self-supervision can be used to learn accurate depth and ego-motion estimation without prior knowledge of the camera model.
Inspired by the geometric model of Grossberg and Nayar, we introduce Neural Ray Surfaces (NRS), convolutional networks that represent pixel-wise projection rays.
We demonstrate the use of NRS for self-supervised learning of visual odometry and depth estimation from raw videos obtained using a wide variety of camera systems.
arXiv Detail & Related papers (2020-08-15T02:29:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.