Deep Multi-view Depth Estimation with Predicted Uncertainty
- URL: http://arxiv.org/abs/2011.09594v2
- Date: Sat, 27 Mar 2021 14:33:40 GMT
- Title: Deep Multi-view Depth Estimation with Predicted Uncertainty
- Authors: Tong Ke, Tien Do, Khiem Vuong, Kourosh Sartipi, and Stergios I.
Roumeliotis
- Abstract summary: We employ a dense-optical-flow network to compute correspondences and then triangulate the point cloud to obtain an initial depth map.
To further increase the triangulation accuracy, we introduce a depth-refinement network (DRN) that optimize the initial depth map based on the image's contextual cues.
- Score: 11.012201499666503
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the problem of estimating dense depth from a
sequence of images using deep neural networks. Specifically, we employ a
dense-optical-flow network to compute correspondences and then triangulate the
point cloud to obtain an initial depth map.Parts of the point cloud, however,
may be less accurate than others due to lack of common observations or small
parallax. To further increase the triangulation accuracy, we introduce a
depth-refinement network (DRN) that optimizes the initial depth map based on
the image's contextual cues. In particular, the DRN contains an iterative
refinement module (IRM) that improves the depth accuracy over iterations by
refining the deep features. Lastly, the DRN also predicts the uncertainty in
the refined depths, which is desirable in applications such as measurement
selection for scene reconstruction. We show experimentally that our algorithm
outperforms state-of-the-art approaches in terms of depth accuracy, and verify
that our predicted uncertainty is highly correlated to the actual depth error.
Related papers
- Depth-guided NeRF Training via Earth Mover's Distance [0.6749750044497732]
We propose a novel approach to uncertainty in depth priors for NeRF supervision.
We use off-the-shelf pretrained diffusion models to predict depth and capture uncertainty during the denoising process.
Our depth-guided NeRF outperforms all baselines on standard depth metrics by a large margin.
arXiv Detail & Related papers (2024-03-19T23:54:07Z) - Depth Estimation from Single-shot Monocular Endoscope Image Using Image
Domain Adaptation And Edge-Aware Depth Estimation [1.7086737326992167]
We propose a depth estimation method from a single-shot monocular endoscopic image using Lambertian surface translation by domain adaptation and depth estimation using multi-scale edge loss.
The texture and specular reflection on the surface of an organ reduce the accuracy of depth estimations.
We applied the estimated depth images to automated anatomical location identification of colonoscopic images using a convolutional neural network.
arXiv Detail & Related papers (2022-01-12T14:06:54Z) - Depth Refinement for Improved Stereo Reconstruction [13.941756438712382]
Current techniques for depth estimation from stereoscopic images still suffer from a built-in drawback.
A simple analysis reveals that the depth error is quadratically proportional to the object's distance.
We propose a simple but effective method that uses a refinement network for depth estimation.
arXiv Detail & Related papers (2021-12-15T12:21:08Z) - Monocular Depth Estimation Primed by Salient Point Detection and
Normalized Hessian Loss [43.950140695759764]
We propose an accurate and lightweight framework for monocular depth estimation based on a self-attention mechanism stemming from salient point detection.
We introduce a normalized Hessian loss term invariant to scaling and shear along the depth direction, which is shown to substantially improve the accuracy.
The proposed method achieves state-of-the-art results on NYU-Depth-v2 and KITTI while using 3.1-38.4 times smaller model in terms of the number of parameters than baseline approaches.
arXiv Detail & Related papers (2021-08-25T07:51:09Z) - Depth Completion using Plane-Residual Representation [84.63079529738924]
We introduce a novel way of interpreting depth information with the closest depth plane label $p$ and a residual value $r$, as we call it, Plane-Residual (PR) representation.
By interpreting depth information in PR representation and using our corresponding depth completion network, we were able to acquire improved depth completion performance with faster computation.
arXiv Detail & Related papers (2021-04-15T10:17:53Z) - PLADE-Net: Towards Pixel-Level Accuracy for Self-Supervised Single-View
Depth Estimation with Neural Positional Encoding and Distilled Matting Loss [49.66736599668501]
We propose a self-supervised single-view pixel-level accurate depth estimation network, called PLADE-Net.
Our method shows unprecedented accuracy levels, exceeding 95% in terms of the $delta1$ metric on the KITTI dataset.
arXiv Detail & Related papers (2021-03-12T15:54:46Z) - CodeVIO: Visual-Inertial Odometry with Learned Optimizable Dense Depth [83.77839773394106]
We present a lightweight, tightly-coupled deep depth network and visual-inertial odometry system.
We provide the network with previously marginalized sparse features from VIO to increase the accuracy of initial depth prediction.
We show that it can run in real-time with single-thread execution while utilizing GPU acceleration only for the network and code Jacobian.
arXiv Detail & Related papers (2020-12-18T09:42:54Z) - Direct Depth Learning Network for Stereo Matching [79.3665881702387]
A novel Direct Depth Learning Network (DDL-Net) is designed for stereo matching.
DDL-Net consists of two stages: the Coarse Depth Estimation stage and the Adaptive-Grained Depth Refinement stage.
We show that DDL-Net achieves an average improvement of 25% on the SceneFlow dataset and $12%$ on the DrivingStereo dataset.
arXiv Detail & Related papers (2020-12-10T10:33:57Z) - Guiding Monocular Depth Estimation Using Depth-Attention Volume [38.92495189498365]
We propose guiding depth estimation to favor planar structures that are ubiquitous especially in indoor environments.
Experiments on two popular indoor datasets, NYU-Depth-v2 and ScanNet, show that our method achieves state-of-the-art depth estimation results.
arXiv Detail & Related papers (2020-04-06T15:45:52Z) - Occlusion-Aware Depth Estimation with Adaptive Normal Constraints [85.44842683936471]
We present a new learning-based method for multi-frame depth estimation from a color video.
Our method outperforms the state-of-the-art in terms of depth estimation accuracy.
arXiv Detail & Related papers (2020-04-02T07:10:45Z) - Depth Completion Using a View-constrained Deep Prior [73.21559000917554]
Recent work has shown that the structure of convolutional neural networks (CNNs) induces a strong prior that favors natural images.
This prior, known as a deep image prior (DIP), is an effective regularizer in inverse problems such as image denoising and inpainting.
We extend the concept of the DIP to depth images. Given color images and noisy and incomplete target depth maps, we reconstruct a depth map restored by virtue of using the CNN network structure as a prior.
arXiv Detail & Related papers (2020-01-21T21:56:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.