Blind Non-Uniform Motion Deblurring using Atrous Spatial Pyramid
Deformable Convolution and Deblurring-Reblurring Consistency
- URL: http://arxiv.org/abs/2106.14336v1
- Date: Sun, 27 Jun 2021 23:14:52 GMT
- Title: Blind Non-Uniform Motion Deblurring using Atrous Spatial Pyramid
Deformable Convolution and Deblurring-Reblurring Consistency
- Authors: Dong Huo, Abbas Masoumzadeh, Yee-Hong Yang
- Abstract summary: We propose a new architecture which consists of multiple Atrous Spatial Pyramid Deformable Convolution modules.
Multiple ASPDC modules implicitly learn the pixel-specific motion with different dilation rates in the same layer to handle movements of different magnitude.
Our experimental results show that the proposed method outperforms state-of-the-art methods on the benchmark datasets.
- Score: 5.994412766684843
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Many deep learning based methods are designed to remove non-uniform
(spatially variant) motion blur caused by object motion and camera shake
without knowing the blur kernel. Some methods directly output the latent sharp
image in one stage, while others utilize a multi-stage strategy (\eg
multi-scale, multi-patch, or multi-temporal) to gradually restore the sharp
image. However, these methods have the following two main issues: 1) The
computational cost of multi-stage is high; 2) The same convolution kernel is
applied in different regions, which is not an ideal choice for non-uniform
blur. Hence, non-uniform motion deblurring is still a challenging and open
problem. In this paper, we propose a new architecture which consists of
multiple Atrous Spatial Pyramid Deformable Convolution (ASPDC) modules to
deblur an image end-to-end with more flexibility. Multiple ASPDC modules
implicitly learn the pixel-specific motion with different dilation rates in the
same layer to handle movements of different magnitude. To improve the training,
we also propose a reblurring network to map the deblurred output back to the
blurred input, which constrains the solution space. Our experimental results
show that the proposed method outperforms state-of-the-art methods on the
benchmark datasets.
Related papers
- Motion-adaptive Separable Collaborative Filters for Blind Motion Deblurring [71.60457491155451]
Eliminating image blur produced by various kinds of motion has been a challenging problem.
We propose a novel real-world deblurring filtering model called the Motion-adaptive Separable Collaborative Filter.
Our method provides an effective solution for real-world motion blur removal and achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-04-19T19:44:24Z) - Efficient Multi-scale Network with Learnable Discrete Wavelet Transform for Blind Motion Deblurring [25.36888929483233]
We propose a multi-scale network based on single-input and multiple-outputs(SIMO) for motion deblurring.
We combine the characteristics of real-world trajectories with a learnable wavelet transform module to focus on the directional continuity and frequency features of the step-by-step transitions between blurred images to sharp images.
arXiv Detail & Related papers (2023-12-29T02:59:40Z) - A Constrained Deformable Convolutional Network for Efficient Single
Image Dynamic Scene Blind Deblurring with Spatially-Variant Motion Blur
Kernels Estimation [12.744989551644744]
We propose a novel constrained deformable convolutional network (CDCN) for efficient single image dynamic scene blind deblurring.
CDCN simultaneously achieves accurate spatially-variant motion blur kernels estimation and the high-quality image restoration.
arXiv Detail & Related papers (2022-08-23T03:28:21Z) - Animation from Blur: Multi-modal Blur Decomposition with Motion Guidance [83.25826307000717]
We study the challenging problem of recovering detailed motion from a single motion-red image.
Existing solutions to this problem estimate a single image sequence without considering the motion ambiguity for each region.
In this paper, we explicitly account for such motion ambiguity, allowing us to generate multiple plausible solutions all in sharp detail.
arXiv Detail & Related papers (2022-07-20T18:05:53Z) - Learning to Estimate Hidden Motions with Global Motion Aggregation [71.12650817490318]
Occlusions pose a significant challenge to optical flow algorithms that rely on local evidences.
We introduce a global motion aggregation module to find long-range dependencies between pixels in the first image.
We demonstrate that the optical flow estimates in the occluded regions can be significantly improved without damaging the performance in non-occluded regions.
arXiv Detail & Related papers (2021-04-06T10:32:03Z) - Single Image Non-uniform Blur Kernel Estimation via Adaptive Basis
Decomposition [1.854931308524932]
We propose a general, non-parametric model for dense non-uniform motion blur estimation.
We show that our method overcomes the limitations of existing non-uniform motion blur estimation.
arXiv Detail & Related papers (2021-02-01T18:02:31Z) - Recurrent Multi-view Alignment Network for Unsupervised Surface
Registration [79.72086524370819]
Learning non-rigid registration in an end-to-end manner is challenging due to the inherent high degrees of freedom and the lack of labeled training data.
We propose to represent the non-rigid transformation with a point-wise combination of several rigid transformations.
We also introduce a differentiable loss function that measures the 3D shape similarity on the projected multi-view 2D depth images.
arXiv Detail & Related papers (2020-11-24T14:22:42Z) - Monocular, One-stage, Regression of Multiple 3D People [105.3143785498094]
We propose to Regress all meshes in a One-stage fashion for Multiple 3D People (termed ROMP)
Our method simultaneously predicts a Body Center heatmap and a Mesh map, which can jointly describe the 3D body mesh on the pixel level.
Compared with state-of-the-art methods, ROMP superior performance on the challenging multi-person benchmarks.
arXiv Detail & Related papers (2020-08-27T17:21:47Z) - MuCAN: Multi-Correspondence Aggregation Network for Video
Super-Resolution [63.02785017714131]
Video super-resolution (VSR) aims to utilize multiple low-resolution frames to generate a high-resolution prediction for each frame.
Inter- and intra-frames are the key sources for exploiting temporal and spatial information.
We build an effective multi-correspondence aggregation network (MuCAN) for VSR.
arXiv Detail & Related papers (2020-07-23T05:41:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.