Learning to Estimate Kernel Scale and Orientation of Defocus Blur with
Asymmetric Coded Aperture
- URL: http://arxiv.org/abs/2103.05843v1
- Date: Wed, 10 Mar 2021 03:12:15 GMT
- Title: Learning to Estimate Kernel Scale and Orientation of Defocus Blur with
Asymmetric Coded Aperture
- Authors: Jisheng Li, Qi Dai, Jiangtao Wen
- Abstract summary: Consistent in-focus input imagery is an essential precondition for machine vision systems to perceive the dynamic environment.
A defocus blur severely degrades the performance of vision systems.
We propose a deep-learning-based framework estimating the kernel scale and orientation of the defocus blur to adjust lens focus rapidly.
- Score: 19.472377706422474
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Consistent in-focus input imagery is an essential precondition for machine
vision systems to perceive the dynamic environment. A defocus blur severely
degrades the performance of vision systems. To tackle this problem, we propose
a deep-learning-based framework estimating the kernel scale and orientation of
the defocus blur to adjust lens focus rapidly. Our pipeline utilizes 3D ConvNet
for a variable number of input hypotheses to select the optimal slice from the
input stack. We use random shuffle and Gumbel-softmax to improve network
performance. We also propose to generate synthetic defocused images with
various asymmetric coded apertures to facilitate training. Experiments are
conducted to demonstrate the effectiveness of our framework.
Related papers
- Reblurring-Guided Single Image Defocus Deblurring: A Learning Framework with Misaligned Training Pairs [65.25002116216771]
We introduce a reblurring-guided learning framework for single image defocus deblurring.
Our reblurring module ensures spatial consistency between the deblurred image, the reblurred image and the input blurry image.
We have collected a new dataset specifically for single image defocus deblurring with typical misalignments.
arXiv Detail & Related papers (2024-09-26T12:37:50Z) - Aberration-Aware Depth-from-Focus [20.956132508261664]
We investigate the domain gap caused by off-axis aberrations that will affect the decision of the best-focused frame in a focal stack.
We then explore bridging this domain gap through aberration-aware training (AAT)
Our approach involves a lightweight network that models lens aberrations at different positions and focus distances, which is then integrated into the conventional network training pipeline.
arXiv Detail & Related papers (2023-03-08T15:21:33Z) - USegScene: Unsupervised Learning of Depth, Optical Flow and Ego-Motion
with Semantic Guidance and Coupled Networks [31.600708674008384]
USegScene is a framework for semantically guided unsupervised learning of depth, optical flow and ego-motion estimation for stereo camera images.
We present results on the popular KITTI dataset and show that our approach outperforms other methods by a large margin.
arXiv Detail & Related papers (2022-07-15T13:25:47Z) - Single-image Defocus Deblurring by Integration of Defocus Map Prediction
Tracing the Inverse Problem Computation [25.438654895178686]
We propose a simple but effective network with spatial modulation based on the defocus map.
Experimental results show that our method can achieve better quantitative and qualitative evaluation performance than the existing state-of-the-art methods.
arXiv Detail & Related papers (2022-07-07T02:15:33Z) - Deep Depth from Focus with Differential Focus Volume [17.505649653615123]
We propose a convolutional neural network (CNN) to find the best-focused pixels in a focal stack and infer depth from the focus estimation.
The key innovation of the network is the novel deep differential focus volume (DFV)
arXiv Detail & Related papers (2021-12-03T04:49:51Z) - Decentralized Autofocusing System with Hierarchical Agents [2.7716102039510564]
We propose a hierarchical multi-agent deep reinforcement learning approach for intelligently controlling the camera and the lens focusing settings.
The algorithm relies on the latent representation of the camera's stream and, thus, it is the first method to allow a completely no-reference tuning of the camera.
arXiv Detail & Related papers (2021-08-29T13:45:15Z) - Single image deep defocus estimation and its applications [82.93345261434943]
We train a deep neural network to classify image patches into one of the 20 levels of blurriness.
The trained model is used to determine the patch blurriness which is then refined by applying an iterative weighted guided filter.
The result is a defocus map that carries the information of the degree of blurriness for each pixel.
arXiv Detail & Related papers (2021-07-30T06:18:16Z) - Universal and Flexible Optical Aberration Correction Using Deep-Prior
Based Deconvolution [51.274657266928315]
We propose a PSF aware plug-and-play deep network, which takes the aberrant image and PSF map as input and produces the latent high quality version via incorporating lens-specific deep priors.
Specifically, we pre-train a base model from a set of diverse lenses and then adapt it to a given lens by quickly refining the parameters.
arXiv Detail & Related papers (2021-04-07T12:00:38Z) - Optical Flow Estimation from a Single Motion-blurred Image [66.2061278123057]
Motion blur in an image may have practical interests in fundamental computer vision problems.
We propose a novel framework to estimate optical flow from a single motion-blurred image in an end-to-end manner.
arXiv Detail & Related papers (2021-03-04T12:45:18Z) - A Flexible Framework for Designing Trainable Priors with Adaptive
Smoothing and Game Encoding [57.1077544780653]
We introduce a general framework for designing and training neural network layers whose forward passes can be interpreted as solving non-smooth convex optimization problems.
We focus on convex games, solved by local agents represented by the nodes of a graph and interacting through regularization functions.
This approach is appealing for solving imaging problems, as it allows the use of classical image priors within deep models that are trainable end to end.
arXiv Detail & Related papers (2020-06-26T08:34:54Z) - Rapid Whole Slide Imaging via Learning-based Two-shot Virtual
Autofocusing [57.90239401665367]
Whole slide imaging (WSI) is an emerging technology for digital pathology.
We propose the concept of textitvirtual autofocusing, which does not rely on mechanical adjustment to conduct refocusing.
arXiv Detail & Related papers (2020-03-14T13:40:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.