Unsupervised Domain-Specific Deblurring using Scale-Specific Attention
- URL: http://arxiv.org/abs/2112.06175v1
- Date: Sun, 12 Dec 2021 07:47:45 GMT
- Title: Unsupervised Domain-Specific Deblurring using Scale-Specific Attention
- Authors: Praveen Kandula and Rajagopalan. A. N
- Abstract summary: We propose unsupervised domain-specific deblurring using a scale-adaptive attention module (SAAM)
Our network does not require supervised pairs for training, and the deblurring mechanism is primarily guided by adversarial loss.
Different ablation studies show that our coarse-to-fine mechanism outperforms end-to-end unsupervised models and SAAM is able to attend better compared to attention models used in literature.
- Score: 0.25797036386508543
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the literature, coarse-to-fine or scale-recurrent approach i.e.
progressively restoring a clean image from its low-resolution versions has been
successfully employed for single image deblurring. However, a major
disadvantage of existing methods is the need for paired data; i.e. sharpblur
image pairs of the same scene, which is a complicated and cumbersome
acquisition procedure. Additionally, due to strong supervision on loss
functions, pre-trained models of such networks are strongly biased towards the
blur experienced during training and tend to give sub-optimal performance when
confronted by new blur kernels during inference time. To address the above
issues, we propose unsupervised domain-specific deblurring using a
scale-adaptive attention module (SAAM). Our network does not require supervised
pairs for training, and the deblurring mechanism is primarily guided by
adversarial loss, thus making our network suitable for a distribution of blur
functions. Given a blurred input image, different resolutions of the same image
are used in our model during training and SAAM allows for effective flow of
information across the resolutions. For network training at a specific scale,
SAAM attends to lower scale features as a function of the current scale.
Different ablation studies show that our coarse-to-fine mechanism outperforms
end-to-end unsupervised models and SAAM is able to attend better compared to
attention models used in literature. Qualitative and quantitative comparisons
(on no-reference metrics) show that our method outperforms prior unsupervised
methods.
Related papers
- Unsupervised Representation Learning by Balanced Self Attention Matching [2.3020018305241337]
We present a self-supervised method for embedding image features called BAM.
We obtain rich representations and avoid feature collapse by minimizing a loss that matches these distributions to their globally balanced and entropy regularized version.
We show competitive performance with leading methods on both semi-supervised and transfer-learning benchmarks.
arXiv Detail & Related papers (2024-08-04T12:52:44Z) - AMSA-UNet: An Asymmetric Multiple Scales U-net Based on Self-attention for Deblurring [7.00986132499006]
asymmetric multiple scales U-net based on self-attention (AMSA-UNet) is proposed to improve the accuracy and computational complexity.
By introducing a multiple-scales U shape architecture, the network can focus on blurry regions at the global level and better recover image details at the local level.
arXiv Detail & Related papers (2024-06-13T11:39:02Z) - Intra-task Mutual Attention based Vision Transformer for Few-Shot Learning [12.5354658533836]
Humans possess remarkable ability to accurately classify new, unseen images after being exposed to only a few examples.
For artificial neural network models, determining the most relevant features for distinguishing between two images with limited samples presents a challenge.
We propose an intra-task mutual attention method for few-shot learning, that involves splitting the support and query samples into patches.
arXiv Detail & Related papers (2024-05-06T02:02:57Z) - Learning to Mask and Permute Visual Tokens for Vision Transformer
Pre-Training [59.923672191632065]
We propose a new self-supervised pre-training approach, named Masked and Permuted Vision Transformer (MaPeT)
MaPeT employs autoregressive and permuted predictions to capture intra-patch dependencies.
Our results demonstrate that MaPeT achieves competitive performance on ImageNet.
arXiv Detail & Related papers (2023-06-12T18:12:19Z) - Masked Images Are Counterfactual Samples for Robust Fine-tuning [77.82348472169335]
Fine-tuning deep learning models can lead to a trade-off between in-distribution (ID) performance and out-of-distribution (OOD) robustness.
We propose a novel fine-tuning method, which uses masked images as counterfactual samples that help improve the robustness of the fine-tuning model.
arXiv Detail & Related papers (2023-03-06T11:51:28Z) - Multiscale Structure Guided Diffusion for Image Deblurring [24.09642909404091]
Diffusion Probabilistic Models (DPMs) have been employed for image deblurring.
We introduce a simple yet effective multiscale structure guidance as an implicit bias.
We demonstrate more robust deblurring results with fewer artifacts on unseen data.
arXiv Detail & Related papers (2022-12-04T10:40:35Z) - LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of
Feature Similarity [49.84167231111667]
Existing works in self-supervised landmark detection are based on learning dense (pixel-level) feature representations from an image.
We introduce an approach to enhance the learning of dense equivariant representations in a self-supervised fashion.
We show that having such a prior in the feature extractor helps in landmark detection, even under drastically limited number of annotations.
arXiv Detail & Related papers (2022-04-06T17:48:18Z) - Denoising Diffusion Restoration Models [110.1244240726802]
Denoising Diffusion Restoration Models (DDRM) is an efficient, unsupervised posterior sampling method.
We demonstrate DDRM's versatility on several image datasets for super-resolution, deblurring, inpainting, and colorization.
arXiv Detail & Related papers (2022-01-27T20:19:07Z) - Self-Adversarial Training incorporating Forgery Attention for Image
Forgery Localization [40.622844703837046]
We propose a self-adversarial training strategy that expands training data dynamically to achieve more robust performance.
We exploit a coarse-to-fine network to enhance the noise inconsistency between original and tampered regions.
Our proposed algorithm steadily outperforms state-of-the-art methods by a clear margin in different benchmark datasets.
arXiv Detail & Related papers (2021-07-06T07:20:08Z) - Unsupervised Learning of Visual Features by Contrasting Cluster
Assignments [57.33699905852397]
We propose an online algorithm, SwAV, that takes advantage of contrastive methods without requiring to compute pairwise comparisons.
Our method simultaneously clusters the data while enforcing consistency between cluster assignments.
Our method can be trained with large and small batches and can scale to unlimited amounts of data.
arXiv Detail & Related papers (2020-06-17T14:00:42Z) - ADRN: Attention-based Deep Residual Network for Hyperspectral Image
Denoising [52.01041506447195]
We propose an attention-based deep residual network to learn a mapping from noisy HSI to the clean one.
Experimental results demonstrate that our proposed ADRN scheme outperforms the state-of-the-art methods both in quantitative and visual evaluations.
arXiv Detail & Related papers (2020-03-04T08:36:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.