Single Image Depth Estimation Trained via Depth from Defocus Cues
- URL: http://arxiv.org/abs/2001.05036v1
- Date: Tue, 14 Jan 2020 20:22:54 GMT
- Title: Single Image Depth Estimation Trained via Depth from Defocus Cues
- Authors: Shir Gur, Lior Wolf
- Abstract summary: Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
- Score: 105.67073923825842
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Estimating depth from a single RGB images is a fundamental task in computer
vision, which is most directly solved using supervised deep learning. In the
field of unsupervised learning of depth from a single RGB image, depth is not
given explicitly. Existing work in the field receives either a stereo pair, a
monocular video, or multiple views, and, using losses that are based on
structure-from-motion, trains a depth estimation network. In this work, we
rely, instead of different views, on depth from focus cues. Learning is based
on a novel Point Spread Function convolutional layer, which applies location
specific kernels that arise from the Circle-Of-Confusion in each image
location. We evaluate our method on data derived from five common datasets for
depth estimation and lightfield images, and present results that are on par
with supervised methods on KITTI and Make3D datasets and outperform
unsupervised learning approaches. Since the phenomenon of depth from defocus is
not dataset specific, we hypothesize that learning based on it would overfit
less to the specific content in each dataset. Our experiments show that this is
indeed the case, and an estimator learned on one dataset using our method
provides better results on other datasets, than the directly supervised
methods.
Related papers
- SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - Multi-task Learning for Monocular Depth and Defocus Estimations with
Real Images [3.682618267671887]
Most existing methods treat depth estimation and defocus estimation as two separate tasks, ignoring the strong connection between them.
We propose a multi-task learning network consisting of an encoder with two decoders to estimate the depth and defocus map from a single focused image.
Our depth and defocus estimations achieve significantly better performance than other state-of-art algorithms.
arXiv Detail & Related papers (2022-08-21T08:59:56Z) - Monocular Depth Estimation Using Cues Inspired by Biological Vision
Systems [22.539300644593936]
Monocular depth estimation (MDE) aims to transform an RGB image of a scene into a pixelwise depth map from the same camera view.
Part of the MDE task is to learn which visual cues in the image can be used for depth estimation, and how.
We demonstrate that explicitly injecting visual cue information into the model is beneficial for depth estimation.
arXiv Detail & Related papers (2022-04-21T19:42:36Z) - P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior [133.76192155312182]
We propose a method that learns to selectively leverage information from coplanar pixels to improve the predicted depth.
An extensive evaluation of our method shows that we set the new state of the art in supervised monocular depth estimation.
arXiv Detail & Related papers (2022-04-05T10:03:52Z) - 360 Depth Estimation in the Wild -- The Depth360 Dataset and the SegFuse
Network [35.03201732370496]
Single-view depth estimation from omnidirectional images has gained popularity with its wide range of applications such as autonomous driving and scene reconstruction.
In this work, we first establish a large-scale dataset with varied settings called Depth360 to tackle the training data problem.
We then propose an end-to-end two-branch multi-task learning network, SegFuse, that mimics the human eye to effectively learn from the dataset.
arXiv Detail & Related papers (2022-02-16T11:56:31Z) - Learning Multi-modal Information for Robust Light Field Depth Estimation [32.64928379844675]
Existing learning-based depth estimation methods from the focal stack lead to suboptimal performance because of the defocus blur.
We propose a multi-modal learning method for robust light field depth estimation.
Our method achieves superior performance than existing representative methods on two light field datasets.
arXiv Detail & Related papers (2021-04-13T06:51:27Z) - Dual Pixel Exploration: Simultaneous Depth Estimation and Image
Restoration [77.1056200937214]
We study the formation of the DP pair which links the blur and the depth information.
We propose an end-to-end DDDNet (DP-based Depth and De Network) to jointly estimate the depth and restore the image.
arXiv Detail & Related papers (2020-12-01T06:53:57Z) - Learning a Geometric Representation for Data-Efficient Depth Estimation
via Gradient Field and Contrastive Loss [29.798579906253696]
We propose a gradient-based self-supervised learning algorithm with momentum contrastive loss to help ConvNets extract the geometric information with unlabeled images.
Our method outperforms the previous state-of-the-art self-supervised learning algorithms and shows the efficiency of labeled data in triple.
arXiv Detail & Related papers (2020-11-06T06:47:19Z) - Defocus Blur Detection via Depth Distillation [64.78779830554731]
We introduce depth information into DBD for the first time.
In detail, we learn the defocus blur from ground truth and the depth distilled from a well-trained depth estimation network.
Our approach outperforms 11 other state-of-the-art methods on two popular datasets.
arXiv Detail & Related papers (2020-07-16T04:58:09Z) - DiverseDepth: Affine-invariant Depth Prediction Using Diverse Data [110.29043712400912]
We present a method for depth estimation with monocular images, which can predict high-quality depth on diverse scenes up to an affine transformation.
Experiments show that our method outperforms previous methods on 8 datasets by a large margin with the zero-shot test setting.
arXiv Detail & Related papers (2020-02-03T05:38:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.