Deep Multi-Scale Feature Learning for Defocus Blur Estimation
- URL: http://arxiv.org/abs/2009.11939v2
- Date: Sun, 7 Nov 2021 17:29:16 GMT
- Title: Deep Multi-Scale Feature Learning for Defocus Blur Estimation
- Authors: Ali Karaali, Naomi Harte, Claudio Rosito Jung
- Abstract summary: This paper presents an edge-based defocus blur estimation method from a single defocused image.
We first distinguish edges that lie at depth discontinuities (called depth edges, for which the blur estimate is ambiguous) from edges that lie at approximately constant depth regions (called pattern edges, for which the blur estimate is well-defined).
We estimate the defocus blur amount at pattern edges only, and explore an scheme based on guided filters that prevents data propagation across the detected depth edges to obtain a dense blur map with well-defined object boundaries.
- Score: 10.455763145066168
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents an edge-based defocus blur estimation method from a
single defocused image. We first distinguish edges that lie at depth
discontinuities (called depth edges, for which the blur estimate is ambiguous)
from edges that lie at approximately constant depth regions (called pattern
edges, for which the blur estimate is well-defined). Then, we estimate the
defocus blur amount at pattern edges only, and explore an interpolation scheme
based on guided filters that prevents data propagation across the detected
depth edges to obtain a dense blur map with well-defined object boundaries.
Both tasks (edge classification and blur estimation) are performed by deep
convolutional neural networks (CNNs) that share weights to learn meaningful
local features from multi-scale patches centered at edge locations. Experiments
on naturally defocused images show that the proposed method presents
qualitative and quantitative results that outperform state-of-the-art (SOTA)
methods, with a good compromise between running time and accuracy.
Related papers
- Blur aware metric depth estimation with multi-focus plenoptic cameras [8.508198765617196]
We present a new metric depth estimation algorithm using only raw images from a multi-focus plenoptic camera.
The proposed approach is especially suited for the multi-focus configuration where several micro-lenses with different focal lengths are used.
arXiv Detail & Related papers (2023-08-08T13:38:50Z) - Depth and DOF Cues Make A Better Defocus Blur Detector [27.33757097343283]
Defocus blur detection (DBD) separates in-focus and out-of-focus regions in an image.
Previous approaches mistakenly mistook homogeneous areas in focus for defocus blur regions.
We propose an approach called D-DFFNet, which incorporates depth and DOF cues in an implicit manner.
arXiv Detail & Related papers (2023-06-20T07:03:37Z) - Fully Self-Supervised Depth Estimation from Defocus Clue [79.63579768496159]
We propose a self-supervised framework that estimates depth purely from a sparse focal stack.
We show that our framework circumvents the needs for the depth and AIF image ground-truth, and receives superior predictions.
arXiv Detail & Related papers (2023-03-19T19:59:48Z) - Single image deep defocus estimation and its applications [82.93345261434943]
We train a deep neural network to classify image patches into one of the 20 levels of blurriness.
The trained model is used to determine the patch blurriness which is then refined by applying an iterative weighted guided filter.
The result is a defocus map that carries the information of the degree of blurriness for each pixel.
arXiv Detail & Related papers (2021-07-30T06:18:16Z) - Edge-aware Bidirectional Diffusion for Dense Depth Estimation from Light
Fields [31.941861222005603]
We present an algorithm to estimate fast and accurate depth maps from light fields via a sparse set of depth edges and gradients.
Our proposed approach is based around the idea that true depth edges are more sensitive than texture edges to local constraints.
arXiv Detail & Related papers (2021-07-07T01:26:25Z) - Adaptive confidence thresholding for monocular depth estimation [83.06265443599521]
We propose a new approach to leverage pseudo ground truth depth maps of stereo images generated from self-supervised stereo matching methods.
The confidence map of the pseudo ground truth depth map is estimated to mitigate performance degeneration by inaccurate pseudo depth maps.
Experimental results demonstrate superior performance to state-of-the-art monocular depth estimation methods.
arXiv Detail & Related papers (2020-09-27T13:26:16Z) - View-consistent 4D Light Field Depth Estimation [37.04038603184669]
We propose a method to compute depth maps for every sub-aperture image in a light field in a view consistent way.
Our method precisely defines depth edges via EPIs, then we diffuse these edges spatially within the central view.
arXiv Detail & Related papers (2020-09-09T01:47:34Z) - Non-Local Spatial Propagation Network for Depth Completion [82.60915972250706]
We propose a robust and efficient end-to-end non-local spatial propagation network for depth completion.
The proposed network takes RGB and sparse depth images as inputs and estimates non-local neighbors and their affinities of each pixel.
We show that the proposed algorithm is superior to conventional algorithms in terms of depth completion accuracy and robustness to the mixed-depth problem.
arXiv Detail & Related papers (2020-07-20T12:26:51Z) - Defocus Blur Detection via Depth Distillation [64.78779830554731]
We introduce depth information into DBD for the first time.
In detail, we learn the defocus blur from ground truth and the depth distilled from a well-trained depth estimation network.
Our approach outperforms 11 other state-of-the-art methods on two popular datasets.
arXiv Detail & Related papers (2020-07-16T04:58:09Z) - Predicting Sharp and Accurate Occlusion Boundaries in Monocular Depth
Estimation Using Displacement Fields [25.3479048674598]
Current methods for depth map prediction from monocular images tend to predict smooth, poorly localized contours.
We learn to predict, given a depth map predicted by some reconstruction method, a 2D displacement field able to re-sample pixels around the occlusion boundaries into sharper reconstructions.
Our method can be applied to the output of any depth estimation method, in an end-to-end trainable fashion.
arXiv Detail & Related papers (2020-02-28T14:15:07Z) - Single Image Depth Estimation Trained via Depth from Defocus Cues [105.67073923825842]
Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
arXiv Detail & Related papers (2020-01-14T20:22:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.