Self-Supervised Light Field Depth Estimation Using Epipolar Plane Images
- URL: http://arxiv.org/abs/2203.15171v1
- Date: Tue, 29 Mar 2022 01:18:59 GMT
- Title: Self-Supervised Light Field Depth Estimation Using Epipolar Plane Images
- Authors: Kunyuan Li, Jun Zhang, Jun Gao, Meibin Qi
- Abstract summary: We propose a self-supervised learning framework for light field depth estimation.
Compared with other state-of-the-art methods, the proposed method can also obtain higher quality results in real-world scenarios.
- Score: 13.137957601685041
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Exploiting light field data makes it possible to obtain dense and accurate
depth map. However, synthetic scenes with limited disparity range cannot
contain the diversity of real scenes. By training in synthetic data, current
learning-based methods do not perform well in real scenes. In this paper, we
propose a self-supervised learning framework for light field depth estimation.
Different from the existing end-to-end training methods using disparity label
per pixel, our approach implements network training by estimating EPI disparity
shift after refocusing, which extends the disparity range of epipolar lines. To
reduce the sensitivity of EPI to noise, we propose a new input mode called
EPI-Stack, which stacks EPIs in the view dimension. This method is less
sensitive to noise scenes than traditional input mode and improves the
efficiency of estimation. Compared with other state-of-the-art methods, the
proposed method can also obtain higher quality results in real-world scenarios,
especially in the complex occlusion and depth discontinuity.
Related papers
- Revisiting Disparity from Dual-Pixel Images: Physics-Informed Lightweight Depth Estimation [3.6337378417255177]
We propose a lightweight disparity estimation method based on a completion-based network.
By modeling the DP-specific disparity error parametrically and using it for sampling during training, the network acquires the unique properties of DP.
As a result, the proposed method achieved state-of-the-art results while reducing the overall system size to 1/5 of that of the conventional method.
arXiv Detail & Related papers (2024-11-06T09:03:53Z) - Robust Depth Enhancement via Polarization Prompt Fusion Tuning [112.88371907047396]
We present a framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors.
Our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors.
To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets.
arXiv Detail & Related papers (2024-04-05T17:55:33Z) - Learning to Adapt CLIP for Few-Shot Monocular Depth Estimation [31.34615135846137]
We propose a few-shot-based method which learns to adapt the Vision-Language Models for monocular depth estimation.
Specifically, it assigns different depth bins for different scenes, which can be selected by the model during inference.
With only one image per scene for training, our extensive experiment results on the NYU V2 and KITTI dataset demonstrate that our method outperforms the previous state-of-the-art method by up to 10.6% in terms of MARE.
arXiv Detail & Related papers (2023-11-02T06:56:50Z) - Single Image Depth Prediction Made Better: A Multivariate Gaussian Take [163.14849753700682]
We introduce an approach that performs continuous modeling of per-pixel depth.
Our method's accuracy (named MG) is among the top on the KITTI depth-prediction benchmark leaderboard.
arXiv Detail & Related papers (2023-03-31T16:01:03Z) - RA-Depth: Resolution Adaptive Self-Supervised Monocular Depth Estimation [27.679479140943503]
We propose a resolution adaptive self-supervised monocular depth estimation method (RA-Depth) by learning the scale invariance of the scene depth.
RA-Depth achieves state-of-the-art performance, and also exhibits a good ability of resolution adaptation.
arXiv Detail & Related papers (2022-07-25T08:49:59Z) - Light Field Depth Estimation via Stitched Epipolar Plane Images [45.5616314866457]
We propose the concept of stitched-EPI (SEPI) to enhance slope computation.
SEPI achieves this by shifting and concatenating lines from different EPIs that correspond to the same 3D point.
We also present a depth propagation strategy aimed at improving depth estimation in texture-less regions.
arXiv Detail & Related papers (2022-03-29T02:43:40Z) - Occlusion-aware Unsupervised Learning of Depth from 4-D Light Fields [50.435129905215284]
We present an unsupervised learning-based depth estimation method for 4-D light field processing and analysis.
Based on the basic knowledge of the unique geometry structure of light field data, we explore the angular coherence among subsets of the light field views to estimate depth maps.
Our method can significantly shrink the performance gap between the previous unsupervised method and supervised ones, and produce depth maps with comparable accuracy to traditional methods with obviously reduced computational cost.
arXiv Detail & Related papers (2021-06-06T06:19:50Z) - An Adaptive Framework for Learning Unsupervised Depth Completion [59.17364202590475]
We present a method to infer a dense depth map from a color image and associated sparse depth measurements.
We show that regularization and co-visibility are related via the fitness of the model to data and can be unified into a single framework.
arXiv Detail & Related papers (2021-06-06T02:27:55Z) - Towards Unpaired Depth Enhancement and Super-Resolution in the Wild [121.96527719530305]
State-of-the-art data-driven methods of depth map super-resolution rely on registered pairs of low- and high-resolution depth maps of the same scenes.
We consider an approach to depth map enhancement based on learning from unpaired data.
arXiv Detail & Related papers (2021-05-25T16:19:16Z) - Single Image Depth Estimation Trained via Depth from Defocus Cues [105.67073923825842]
Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
arXiv Detail & Related papers (2020-01-14T20:22:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.