P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior
- URL: http://arxiv.org/abs/2204.02091v1
- Date: Tue, 5 Apr 2022 10:03:52 GMT
- Title: P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior
- Authors: Vaishakh Patil, Christos Sakaridis, Alexander Liniger, Luc Van Gool
- Abstract summary: We propose a method that learns to selectively leverage information from coplanar pixels to improve the predicted depth.
An extensive evaluation of our method shows that we set the new state of the art in supervised monocular depth estimation.
- Score: 133.76192155312182
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Monocular depth estimation is vital for scene understanding and downstream
tasks. We focus on the supervised setup, in which ground-truth depth is
available only at training time. Based on knowledge about the high regularity
of real 3D scenes, we propose a method that learns to selectively leverage
information from coplanar pixels to improve the predicted depth. In particular,
we introduce a piecewise planarity prior which states that for each pixel,
there is a seed pixel which shares the same planar 3D surface with the former.
Motivated by this prior, we design a network with two heads. The first head
outputs pixel-level plane coefficients, while the second one outputs a dense
offset vector field that identifies the positions of seed pixels. The plane
coefficients of seed pixels are then used to predict depth at each position.
The resulting prediction is adaptively fused with the initial prediction from
the first head via a learned confidence to account for potential deviations
from precise local planarity. The entire architecture is trained end-to-end
thanks to the differentiability of the proposed modules and it learns to
predict regular depth maps, with sharp edges at occlusion boundaries. An
extensive evaluation of our method shows that we set the new state of the art
in supervised monocular depth estimation, surpassing prior methods on NYU
Depth-v2 and on the Garg split of KITTI. Our method delivers depth maps that
yield plausible 3D reconstructions of the input scenes. Code is available at:
https://github.com/SysCV/P3Depth
Related papers
- DoubleTake: Geometry Guided Depth Estimation [17.464549832122714]
Estimating depth from a sequence of posed RGB images is a fundamental computer vision task.
We introduce a reconstruction which combines volume features with a hint of the prior geometry, rendered as a depth map from the current camera location.
We demonstrate that our method can run at interactive speeds, state-of-the-art estimates of depth and 3D scene in both offline and incremental evaluation scenarios.
arXiv Detail & Related papers (2024-06-26T14:29:05Z) - NDDepth: Normal-Distance Assisted Monocular Depth Estimation and
Completion [18.974297347310287]
We introduce novel physics (geometry)-driven deep learning frameworks for monocular depth estimation and completion.
Our method exceeds in performance prior state-of-the-art monocular depth estimation and completion competitors.
arXiv Detail & Related papers (2023-11-13T09:01:50Z) - ARAI-MVSNet: A multi-view stereo depth estimation network with adaptive
depth range and depth interval [19.28042366225802]
Multi-View Stereo(MVS) is a fundamental problem in geometric computer vision.
We present a novel multi-stage coarse-to-fine framework to achieve adaptive all-pixel depth range and depth interval.
Our model achieves state-of-the-art performance and yields competitive generalization ability.
arXiv Detail & Related papers (2023-08-17T14:52:11Z) - Towards Accurate Reconstruction of 3D Scene Shape from A Single
Monocular Image [91.71077190961688]
We propose a two-stage framework that first predicts depth up to an unknown scale and shift from a single monocular image.
We then exploits 3D point cloud data to predict the depth shift and the camera's focal length that allow us to recover 3D scene shapes.
We test our depth model on nine unseen datasets and achieve state-of-the-art performance on zero-shot evaluation.
arXiv Detail & Related papers (2022-08-28T16:20:14Z) - Virtual Normal: Enforcing Geometric Constraints for Accurate and Robust
Depth Prediction [87.08227378010874]
We show the importance of the high-order 3D geometric constraints for depth prediction.
By designing a loss term that enforces a simple geometric constraint, we significantly improve the accuracy and robustness of monocular depth estimation.
We show state-of-the-art results of learning metric depth on NYU Depth-V2 and KITTI.
arXiv Detail & Related papers (2021-03-07T00:08:21Z) - Learning to Recover 3D Scene Shape from a Single Image [98.20106822614392]
We propose a two-stage framework that first predicts depth up to an unknown scale and shift from a single monocular image.
We then use 3D point cloud encoders to predict the missing depth shift and focal length that allow us to recover a realistic 3D scene shape.
arXiv Detail & Related papers (2020-12-17T02:35:13Z) - Occlusion-Aware Depth Estimation with Adaptive Normal Constraints [85.44842683936471]
We present a new learning-based method for multi-frame depth estimation from a color video.
Our method outperforms the state-of-the-art in terms of depth estimation accuracy.
arXiv Detail & Related papers (2020-04-02T07:10:45Z) - DiverseDepth: Affine-invariant Depth Prediction Using Diverse Data [110.29043712400912]
We present a method for depth estimation with monocular images, which can predict high-quality depth on diverse scenes up to an affine transformation.
Experiments show that our method outperforms previous methods on 8 datasets by a large margin with the zero-shot test setting.
arXiv Detail & Related papers (2020-02-03T05:38:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.