Gabor-guided transformer for single image deraining
- URL: http://arxiv.org/abs/2403.07380v1
- Date: Tue, 12 Mar 2024 07:41:51 GMT
- Title: Gabor-guided transformer for single image deraining
- Authors: Sijin He, Guangfeng Lin
- Abstract summary: We propose a Gabor-guided tranformer (Gabformer) for single image deraining.
The focus on local texture features is enhanced by incorporating the information processed by the Gabor filter into the query vector.
Our method outperforms state-of-the-art approaches.
- Score: 2.330361251490783
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image deraining have have gained a great deal of attention in order to
address the challenges posed by the effects of harsh weather conditions on
visual tasks. While convolutional neural networks (CNNs) are popular, their
limitations in capturing global information may result in ineffective rain
removal. Transformer-based methods with self-attention mechanisms have
improved, but they tend to distort high-frequency details that are crucial for
image fidelity. To solve this problem, we propose the Gabor-guided tranformer
(Gabformer) for single image deraining. The focus on local texture features is
enhanced by incorporating the information processed by the Gabor filter into
the query vector, which also improves the robustness of the model to noise due
to the properties of the filter. Extensive experiments on the benchmarks
demonstrate that our method outperforms state-of-the-art approaches.
Related papers
- Improving Image De-raining Using Reference-Guided Transformers [9.867364371892693]
We present a reference-guided de-raining filter, a transformer network that enhances de-raining results using a reference clean image as guidance.
We validate our method on three datasets and show that our module can improve the performance of existing prior-based, CNN-based, and transformer-based approaches.
arXiv Detail & Related papers (2024-08-01T03:31:45Z) - Training Transformer Models by Wavelet Losses Improves Quantitative and Visual Performance in Single Image Super-Resolution [6.367865391518726]
Transformer-based models have achieved remarkable results in low-level vision tasks including image super-resolution (SR)
To activate more input pixels globally, hybrid attention models have been proposed.
We employ wavelet losses to train Transformer models to improve quantitative and subjective performance.
arXiv Detail & Related papers (2024-04-17T11:25:19Z) - Exploiting Regional Information Transformer for Single Image Deraining [40.96287901893822]
Region Transformer Block (RTB) integrates a Region Masked Attention (RMA) mechanism and a Mixed Gate Forward Block (MGFB)
Our model reaches state-of-the-art performance, significantly improving the image deraining quality.
arXiv Detail & Related papers (2024-02-25T09:09:30Z) - WaveletFormerNet: A Transformer-based Wavelet Network for Real-world
Non-homogeneous and Dense Fog Removal [11.757602977709517]
This paper proposes a Transformer-based wavelet network (WaveletFormerNet) for real-world foggy image recovery.
We introduce parallel convolution in the Transformer block, which allows for the capture of multi-frequency information in a lightweight mechanism.
Our experiments demonstrate that our WaveletFormerNet performs better than state-of-the-art methods.
arXiv Detail & Related papers (2024-01-09T13:42:21Z) - DGNet: Dynamic Gradient-Guided Network for Water-Related Optics Image
Enhancement [77.0360085530701]
Underwater image enhancement (UIE) is a challenging task due to the complex degradation caused by underwater environments.
Previous methods often idealize the degradation process, and neglect the impact of medium noise and object motion on the distribution of image features.
Our approach utilizes predicted images to dynamically update pseudo-labels, adding a dynamic gradient to optimize the network's gradient space.
arXiv Detail & Related papers (2023-12-12T06:07:21Z) - Learning Heavily-Degraded Prior for Underwater Object Detection [59.5084433933765]
This paper seeks transferable prior knowledge from detector-friendly images.
It is based on statistical observations that, the heavily degraded regions of detector-friendly (DFUI) and underwater images have evident feature distribution gaps.
Our method with higher speeds and less parameters still performs better than transformer-based detectors.
arXiv Detail & Related papers (2023-08-24T12:32:46Z) - Single Image Deraining via Feature-based Deep Convolutional Neural
Network [13.39233717329633]
A single image deraining algorithm based on the combination of data-driven and model-based approaches is proposed.
Experiments show that the proposed algorithm significantly outperforms state-of-the-art methods in terms of both qualitative and quantitative measures.
arXiv Detail & Related papers (2023-05-03T13:12:51Z) - Image Deblurring by Exploring In-depth Properties of Transformer [86.7039249037193]
We leverage deep features extracted from a pretrained vision transformer (ViT) to encourage recovered images to be sharp without sacrificing the performance measured by the quantitative metrics.
By comparing the transformer features between recovered image and target one, the pretrained transformer provides high-resolution blur-sensitive semantic information.
One regards the features as vectors and computes the discrepancy between representations extracted from recovered image and target one in Euclidean space.
arXiv Detail & Related papers (2023-03-24T14:14:25Z) - Improving robustness against common corruptions with frequency biased
models [112.65717928060195]
unseen image corruptions can cause a surprisingly large drop in performance.
Image corruption types have different characteristics in the frequency spectrum and would benefit from a targeted type of data augmentation.
We propose a new regularization scheme that minimizes the total variation (TV) of convolution feature-maps to increase high-frequency robustness.
arXiv Detail & Related papers (2021-03-30T10:44:50Z) - Invertible Image Rescaling [118.2653765756915]
We develop an Invertible Rescaling Net (IRN) to produce visually-pleasing low-resolution images.
We capture the distribution of the lost information using a latent variable following a specified distribution in the downscaling process.
arXiv Detail & Related papers (2020-05-12T09:55:53Z) - Blur, Noise, and Compression Robust Generative Adversarial Networks [85.68632778835253]
We propose blur, noise, and compression robust GAN (BNCR-GAN) to learn a clean image generator directly from degraded images.
Inspired by NR-GAN, BNCR-GAN uses a multiple-generator model composed of image, blur- Kernel, noise, and quality-factor generators.
We demonstrate the effectiveness of BNCR-GAN through large-scale comparative studies on CIFAR-10 and a generality analysis on FFHQ.
arXiv Detail & Related papers (2020-03-17T17:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.