DRT: A Lightweight Single Image Deraining Recursive Transformer
- URL: http://arxiv.org/abs/2204.11385v1
- Date: Mon, 25 Apr 2022 01:06:09 GMT
- Title: DRT: A Lightweight Single Image Deraining Recursive Transformer
- Authors: Yuanchu Liang, Saeed Anwar, Yang Liu
- Abstract summary: Over parameterization is a common technique in deep learning to help models learn and generalize sufficiently to the given task.
Recent powerful transformer-based deep learning models on vision tasks usually have heavy parameters and bear training difficulty.
We introduce a self-attention structure with residual connections and propose deraining a recursive transformer (DRT)
Our proposed model uses only 1.3% of the number of parameters of the current best performing model in deraining while exceeding the state-of-the-art methods on the Rain100L benchmark by at least 0.33 dB.
- Score: 21.889582347604648
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over parameterization is a common technique in deep learning to help models
learn and generalize sufficiently to the given task; nonetheless, this often
leads to enormous network structures and consumes considerable computing
resources during training. Recent powerful transformer-based deep learning
models on vision tasks usually have heavy parameters and bear training
difficulty. However, many dense-prediction low-level computer vision tasks,
such as rain streak removing, often need to be executed on devices with limited
computing power and memory in practice. Hence, we introduce a recursive local
window-based self-attention structure with residual connections and propose
deraining a recursive transformer (DRT), which enjoys the superiority of the
transformer but requires a small amount of computing resources. In particular,
through recursive architecture, our proposed model uses only 1.3% of the number
of parameters of the current best performing model in deraining while exceeding
the state-of-the-art methods on the Rain100L benchmark by at least 0.33 dB.
Ablation studies also investigate the impact of recursions on derain outcomes.
Moreover, since the model contains no deliberate design for deraining, it can
also be applied to other image restoration tasks. Our experiment shows that it
can achieve competitive results on desnowing. The source code and pretrained
model can be found at https://github.com/YC-Liang/DRT.
Related papers
- Emergent Agentic Transformer from Chain of Hindsight Experience [96.56164427726203]
We show that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
This is the first time that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
arXiv Detail & Related papers (2023-05-26T00:43:02Z) - Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution [91.3781512926942]
Image super-resolution (SR) has witnessed extensive neural network designs from CNN to transformer architectures.
This work investigates the potential of network pruning for super-resolution iteration to take advantage of off-the-shelf network designs and reduce the underlying computational overhead.
We propose a novel Iterative Soft Shrinkage-Percentage (ISS-P) method by optimizing the sparse structure of a randomly network at each and tweaking unimportant weights with a small amount proportional to the magnitude scale on-the-fly.
arXiv Detail & Related papers (2023-03-16T21:06:13Z) - Magic ELF: Image Deraining Meets Association Learning and Transformer [63.761812092934576]
This paper aims to unify CNN and Transformer to take advantage of their learning merits for image deraining.
A novel multi-input attention module (MAM) is proposed to associate rain removal and background recovery.
Our proposed method (dubbed as ELF) outperforms the state-of-the-art approach (MPRNet) by 0.25 dB on average.
arXiv Detail & Related papers (2022-07-21T12:50:54Z) - Continual Learning with Transformers for Image Classification [12.028617058465333]
In computer vision, neural network models struggle to continually learn new concepts without forgetting what has been learnt in the past.
We develop a solution called Adaptive Distillation of Adapters (ADA), which is developed to perform continual learning.
We empirically demonstrate on different classification tasks that this method maintains a good predictive performance without retraining the model.
arXiv Detail & Related papers (2022-06-28T15:30:10Z) - Lightweight Bimodal Network for Single-Image Super-Resolution via
Symmetric CNN and Recursive Transformer [27.51790638626891]
Single-image super-resolution (SISR) has achieved significant breakthroughs with the development of deep learning.
To solve this issue, we propose a Lightweight Bimodal Network (LBNet) for SISR.
Specifically, an effective Symmetric CNN is designed for local feature extraction and coarse image reconstruction.
arXiv Detail & Related papers (2022-04-28T04:43:22Z) - Restormer: Efficient Transformer for High-Resolution Image Restoration [118.9617735769827]
convolutional neural networks (CNNs) perform well at learning generalizable image priors from large-scale data.
Transformers have shown significant performance gains on natural language and high-level vision tasks.
Our model, named Restoration Transformer (Restormer), achieves state-of-the-art results on several image restoration tasks.
arXiv Detail & Related papers (2021-11-18T18:59:10Z) - Sliced Recursive Transformer [23.899076070924153]
Recursive operation on vision transformers can improve parameter utilization without involving additional parameters.
Our model Sliced Recursive Transformer (SReT) is compatible with a broad range of other designs for efficient vision transformers.
arXiv Detail & Related papers (2021-11-09T17:59:14Z) - From Rain Generation to Rain Removal [67.71728610434698]
We build a full Bayesian generative model for rainy image where the rain layer is parameterized as a generator.
We employ the variational inference framework to approximate the expected statistical distribution of rainy image.
Comprehensive experiments substantiate that the proposed model can faithfully extract the complex rain distribution.
arXiv Detail & Related papers (2020-08-08T18:56:51Z) - A Model-driven Deep Neural Network for Single Image Rain Removal [52.787356046951494]
We propose a model-driven deep neural network for the task, with fully interpretable network structures.
Based on the convolutional dictionary learning mechanism for representing rain, we propose a novel single image deraining model.
All the rain kernels and operators can be automatically extracted, faithfully characterizing the features of both rain and clean background layers.
arXiv Detail & Related papers (2020-05-04T09:13:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.