Learning A Sparse Transformer Network for Effective Image Deraining
- URL: http://arxiv.org/abs/2303.11950v1
- Date: Tue, 21 Mar 2023 15:41:57 GMT
- Title: Learning A Sparse Transformer Network for Effective Image Deraining
- Authors: Xiang Chen, Hao Li, Mingqiang Li, Jinshan Pan
- Abstract summary: We propose an effective DeRaining network, Sparse Transformer (DRSformer)
We develop a learnable top-k selection operator to adaptively retain the most crucial attention scores from the keys for each query for better feature aggregation.
We equip our model with mixture of experts feature compensator to present a cooperation refinement deraining scheme.
- Score: 42.01684644627124
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transformers-based methods have achieved significant performance in image
deraining as they can model the non-local information which is vital for
high-quality image reconstruction. In this paper, we find that most existing
Transformers usually use all similarities of the tokens from the query-key
pairs for the feature aggregation. However, if the tokens from the query are
different from those of the key, the self-attention values estimated from these
tokens also involve in feature aggregation, which accordingly interferes with
the clear image restoration. To overcome this problem, we propose an effective
DeRaining network, Sparse Transformer (DRSformer) that can adaptively keep the
most useful self-attention values for feature aggregation so that the
aggregated features better facilitate high-quality image reconstruction.
Specifically, we develop a learnable top-k selection operator to adaptively
retain the most crucial attention scores from the keys for each query for
better feature aggregation. Simultaneously, as the naive feed-forward network
in Transformers does not model the multi-scale information that is important
for latent clear image restoration, we develop an effective mixed-scale
feed-forward network to generate better features for image deraining. To learn
an enriched set of hybrid features, which combines local context from CNN
operators, we equip our model with mixture of experts feature compensator to
present a cooperation refinement deraining scheme. Extensive experimental
results on the commonly used benchmarks demonstrate that the proposed method
achieves favorable performance against state-of-the-art approaches. The source
code and trained models are available at
https://github.com/cschenxiang/DRSformer.
Related papers
- Transformer-based Clipped Contrastive Quantization Learning for
Unsupervised Image Retrieval [15.982022297570108]
Unsupervised image retrieval aims to learn the important visual characteristics without any given level to retrieve the similar images for a given query image.
In this paper, we propose a TransClippedCLR model by encoding the global context of an image using Transformer having local context through patch based processing.
Results using the proposed clipped contrastive learning are greatly improved on all datasets as compared to same backbone network with vanilla contrastive learning.
arXiv Detail & Related papers (2024-01-27T09:39:11Z) - HAT: Hybrid Attention Transformer for Image Restoration [61.74223315807691]
Transformer-based methods have shown impressive performance in image restoration tasks, such as image super-resolution and denoising.
We propose a new Hybrid Attention Transformer (HAT) to activate more input pixels for better restoration.
Our HAT achieves state-of-the-art performance both quantitatively and qualitatively.
arXiv Detail & Related papers (2023-09-11T05:17:55Z) - Prompt-based Ingredient-Oriented All-in-One Image Restoration [0.0]
We propose a novel data ingredient-oriented approach to tackle multiple image degradation tasks.
Specifically, we utilize a encoder to capture features and introduce prompts with degradation-specific information to guide the decoder.
Our method performs competitively to the state-of-the-art.
arXiv Detail & Related papers (2023-09-06T15:05:04Z) - Learning Image Deraining Transformer Network with Dynamic Dual
Self-Attention [46.11162082219387]
This paper proposes an effective image deraining Transformer with dynamic dual self-attention (DDSA)
Specifically, we only select the most useful similarity values based on top-k approximate calculation to achieve sparse attention.
In addition, we also develop a novel spatial-enhanced feed-forward network (SEFN) to further obtain a more accurate representation for achieving high-quality derained results.
arXiv Detail & Related papers (2023-08-15T13:59:47Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Restormer: Efficient Transformer for High-Resolution Image Restoration [118.9617735769827]
convolutional neural networks (CNNs) perform well at learning generalizable image priors from large-scale data.
Transformers have shown significant performance gains on natural language and high-level vision tasks.
Our model, named Restoration Transformer (Restormer), achieves state-of-the-art results on several image restoration tasks.
arXiv Detail & Related papers (2021-11-18T18:59:10Z) - XCiT: Cross-Covariance Image Transformers [73.33400159139708]
We propose a "transposed" version of self-attention that operates across feature channels rather than tokens.
The resulting cross-covariance attention (XCA) has linear complexity in the number of tokens, and allows efficient processing of high-resolution images.
arXiv Detail & Related papers (2021-06-17T17:33:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.