NDDepth: Normal-Distance Assisted Monocular Depth Estimation and
Completion
- URL: http://arxiv.org/abs/2311.07166v1
- Date: Mon, 13 Nov 2023 09:01:50 GMT
- Title: NDDepth: Normal-Distance Assisted Monocular Depth Estimation and
Completion
- Authors: Shuwei Shao, Zhongcai Pei, Weihai Chen, Peter C. Y. Chen and Zhengguo
Li
- Abstract summary: We introduce novel physics (geometry)-driven deep learning frameworks for monocular depth estimation and completion.
Our method exceeds in performance prior state-of-the-art monocular depth estimation and completion competitors.
- Score: 18.974297347310287
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over the past few years, monocular depth estimation and completion have been
paid more and more attention from the computer vision community because of
their widespread applications. In this paper, we introduce novel physics
(geometry)-driven deep learning frameworks for these two tasks by assuming that
3D scenes are constituted with piece-wise planes. Instead of directly
estimating the depth map or completing the sparse depth map, we propose to
estimate the surface normal and plane-to-origin distance maps or complete the
sparse surface normal and distance maps as intermediate outputs. To this end,
we develop a normal-distance head that outputs pixel-level surface normal and
distance. Meanwhile, the surface normal and distance maps are regularized by a
developed plane-aware consistency constraint, which are then transformed into
depth maps. Furthermore, we integrate an additional depth head to strengthen
the robustness of the proposed frameworks. Extensive experiments on the
NYU-Depth-v2, KITTI and SUN RGB-D datasets demonstrate that our method exceeds
in performance prior state-of-the-art monocular depth estimation and completion
competitors. The source code will be available at
https://github.com/ShuweiShao/NDDepth.
Related papers
- Refinement of Monocular Depth Maps via Multi-View Differentiable Rendering [4.717325308876748]
We present a novel approach to generate view consistent and detailed depth maps from a number of posed images.
We leverage advances in monocular depth estimation, which generate topologically complete, but metrically inaccurate depth maps.
Our method is able to generate dense, detailed, high-quality depth maps, also in challenging indoor scenarios, and outperforms state-of-the-art depth reconstruction approaches.
arXiv Detail & Related papers (2024-10-04T18:50:28Z) - NDDepth: Normal-Distance Assisted Monocular Depth Estimation [22.37113584192617]
We propose a novel physics (geometry)-driven deep learning framework for monocular depth estimation.
We introduce a new normal-distance head that outputs pixel-level surface normal and plane-to-origin distance for deriving depth at each position.
We develop an effective contrastive iterative refinement module that refines depth in a complementary manner according to the depth uncertainty.
arXiv Detail & Related papers (2023-09-19T13:05:57Z) - P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior [133.76192155312182]
We propose a method that learns to selectively leverage information from coplanar pixels to improve the predicted depth.
An extensive evaluation of our method shows that we set the new state of the art in supervised monocular depth estimation.
arXiv Detail & Related papers (2022-04-05T10:03:52Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - Virtual Normal: Enforcing Geometric Constraints for Accurate and Robust
Depth Prediction [87.08227378010874]
We show the importance of the high-order 3D geometric constraints for depth prediction.
By designing a loss term that enforces a simple geometric constraint, we significantly improve the accuracy and robustness of monocular depth estimation.
We show state-of-the-art results of learning metric depth on NYU Depth-V2 and KITTI.
arXiv Detail & Related papers (2021-03-07T00:08:21Z) - GeoNet++: Iterative Geometric Neural Network with Edge-Aware Refinement
for Joint Depth and Surface Normal Estimation [204.13451624763735]
We propose a geometric neural network with edge-aware refinement (GeoNet++) to jointly predict both depth and surface normal maps from a single image.
GeoNet++ effectively predicts depth and surface normals with strong 3D consistency and sharp boundaries.
In contrast to current metrics that focus on evaluating pixel-wise error/accuracy, 3DGM measures whether the predicted depth can reconstruct high-quality 3D surface normals.
arXiv Detail & Related papers (2020-12-13T06:48:01Z) - Efficient Depth Completion Using Learned Bases [94.0808155168311]
We propose a new global geometry constraint for depth completion.
By assuming depth maps often lay on low dimensional subspaces, a dense depth map can be approximated by a weighted sum of full-resolution principal depth bases.
arXiv Detail & Related papers (2020-12-02T11:57:37Z) - Attention Aware Cost Volume Pyramid Based Multi-view Stereo Network for
3D Reconstruction [12.728154351588053]
We present an efficient multi-view stereo (MVS) network for 3D reconstruction from multiview images.
We introduce a coarseto-fine depth inference strategy to achieve high resolution depth.
arXiv Detail & Related papers (2020-11-25T13:34:11Z) - Occlusion-Aware Depth Estimation with Adaptive Normal Constraints [85.44842683936471]
We present a new learning-based method for multi-frame depth estimation from a color video.
Our method outperforms the state-of-the-art in terms of depth estimation accuracy.
arXiv Detail & Related papers (2020-04-02T07:10:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.