Depth Estimation from a Single Optical Encoded Image using a Learned
Colored-Coded Aperture
- URL: http://arxiv.org/abs/2309.08033v1
- Date: Thu, 14 Sep 2023 21:30:55 GMT
- Title: Depth Estimation from a Single Optical Encoded Image using a Learned
Colored-Coded Aperture
- Authors: Jhon Lopez, Edwin Vargas, Henry Arguello
- Abstract summary: State-of-the-art approaches improve the discrimination between different depths by introducing a binary-coded aperture (CA) in the lens aperture.
Color-coded apertures (CCA) can also produce color misalignment in the captured image which can be utilized to estimate disparity.
We propose a CCA with a greater number of color filters and richer spectral information to optically encode relevant depth information in a single snapshot.
- Score: 18.830374973687416
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Depth estimation from a single image of a conventional camera is a
challenging task since depth cues are lost during the acquisition process.
State-of-the-art approaches improve the discrimination between different depths
by introducing a binary-coded aperture (CA) in the lens aperture that generates
different coded blur patterns at different depths. Color-coded apertures (CCA)
can also produce color misalignment in the captured image which can be utilized
to estimate disparity. Leveraging advances in deep learning, more recent works
have explored the data-driven design of a diffractive optical element (DOE) for
encoding depth information through chromatic aberrations. However, compared
with binary CA or CCA, DOEs are more expensive to fabricate and require
high-precision devices. Different from previous CCA-based approaches that
employ few basic colors, in this work we propose a CCA with a greater number of
color filters and richer spectral information to optically encode relevant
depth information in a single snapshot. Furthermore, we propose to jointly
learn the color-coded aperture (CCA) pattern and a convolutional neural network
(CNN) to retrieve depth information by using an end-to-end optimization
approach. We demonstrate through different experiments on three different data
sets that the designed color-encoding has the potential to remove depth
ambiguities and provides better depth estimates compared to state-of-the-art
approaches. Additionally, we build a low-cost prototype of our CCA using a
photographic film and validate the proposed approach in real scenarios.
Related papers
- Robust Depth Enhancement via Polarization Prompt Fusion Tuning [112.88371907047396]
We present a framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors.
Our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors.
To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets.
arXiv Detail & Related papers (2024-04-05T17:55:33Z) - Deep Phase Coded Image Prior [34.84063452418995]
Phase-coded imaging is a method to tackle tasks such as passive depth estimation and extended depth of field.
Most of the current deep learning-based methods for depth estimation or all-in-focus imaging require a training dataset with high-quality depth maps.
We propose a new method named "Deep Phase Coded Image Prior" (DPCIP) for jointly recovering the depth map and all-in-focus image.
arXiv Detail & Related papers (2024-04-05T05:58:40Z) - Exploring Deep Learning Image Super-Resolution for Iris Recognition [50.43429968821899]
We propose the use of two deep learning single-image super-resolution approaches: Stacked Auto-Encoders (SAE) and Convolutional Neural Networks (CNN)
We validate the methods with a database of 1.872 near-infrared iris images with quality assessment and recognition experiments showing the superiority of deep learning approaches over the compared algorithms.
arXiv Detail & Related papers (2023-11-02T13:57:48Z) - AGG-Net: Attention Guided Gated-convolutional Network for Depth Image
Completion [1.8820731605557168]
We propose a new model for depth image completion based on the Attention Guided Gated-convolutional Network (AGG-Net)
In the encoding stage, an Attention Guided Gated-Convolution (AG-GConv) module is proposed to realize the fusion of depth and color features at different scales.
In the decoding stage, an Attention Guided Skip Connection (AG-SC) module is presented to avoid introducing too many depth-irrelevant features to the reconstruction.
arXiv Detail & Related papers (2023-09-04T14:16:08Z) - DEHRFormer: Real-time Transformer for Depth Estimation and Haze Removal
from Varicolored Haze Scenes [10.174140482558904]
We propose a real-time transformer for simultaneous single image Depth Estimation and Haze Removal.
DEHRFormer consists of a single encoder and two task-specific decoders.
We introduce a novel learning paradigm that utilizes contrastive learning and domain consistency learning to tackle weak-generalization problem for real-world dehazing.
arXiv Detail & Related papers (2023-03-13T07:47:18Z) - End-to-end Learning for Joint Depth and Image Reconstruction from
Diffracted Rotation [10.896567381206715]
We propose a novel end-to-end learning approach for depth from diffracted rotation.
Our approach requires a significantly less complex model and less training data, yet it is superior to existing methods in the task of monocular depth estimation.
arXiv Detail & Related papers (2022-04-14T16:14:37Z) - PDC: Piecewise Depth Completion utilizing Superpixels [0.0]
Current approaches often rely on CNN-based methods with several known drawbacks.
We propose our novel Piecewise Depth Completion (PDC), which works completely without deep learning.
In our evaluation, we can show both the influence of the individual proposed processing steps and the overall performance of our method on the challenging KITTI dataset.
arXiv Detail & Related papers (2021-07-14T13:58:39Z) - Time-Multiplexed Coded Aperture Imaging: Learned Coded Aperture and
Pixel Exposures for Compressive Imaging Systems [56.154190098338965]
We show that our proposed time multiplexed coded aperture (TMCA) can be optimized end-to-end.
TMCA induces better coded snapshots enabling superior reconstructions in two different applications: compressive light field imaging and hyperspectral imaging.
This codification outperforms the state-of-the-art compressive imaging systems by more than 4dB in those applications.
arXiv Detail & Related papers (2021-04-06T22:42:34Z) - Dual Pixel Exploration: Simultaneous Depth Estimation and Image
Restoration [77.1056200937214]
We study the formation of the DP pair which links the blur and the depth information.
We propose an end-to-end DDDNet (DP-based Depth and De Network) to jointly estimate the depth and restore the image.
arXiv Detail & Related papers (2020-12-01T06:53:57Z) - Defocus Blur Detection via Depth Distillation [64.78779830554731]
We introduce depth information into DBD for the first time.
In detail, we learn the defocus blur from ground truth and the depth distilled from a well-trained depth estimation network.
Our approach outperforms 11 other state-of-the-art methods on two popular datasets.
arXiv Detail & Related papers (2020-07-16T04:58:09Z) - Single Image Depth Estimation Trained via Depth from Defocus Cues [105.67073923825842]
Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
arXiv Detail & Related papers (2020-01-14T20:22:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.