Dual-Path Coupled Image Deraining Network via Spatial-Frequency
Interaction
- URL: http://arxiv.org/abs/2402.04855v1
- Date: Wed, 7 Feb 2024 13:54:15 GMT
- Title: Dual-Path Coupled Image Deraining Network via Spatial-Frequency
Interaction
- Authors: Yuhong He, Aiwen Jiang, Lingfang Jiang, Zhifeng Wang, Lu Wang
- Abstract summary: Existing image deraining methods tend to neglect critical frequency information.
We have developed an innovative Dual-Path Coupled Deraining Network (DPCNet) that integrates information from both spatial and frequency domains.
Our proposed method not only outperforms the existing state-of-the-art deraining method but also achieves visually pleasuring results with excellent robustness on downstream vision tasks.
- Score: 7.682978264249712
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transformers have recently emerged as a significant force in the field of
image deraining. Existing image deraining methods utilize extensive research on
self-attention. Though showcasing impressive results, they tend to neglect
critical frequency information, as self-attention is generally less adept at
capturing high-frequency details. To overcome this shortcoming, we have
developed an innovative Dual-Path Coupled Deraining Network (DPCNet) that
integrates information from both spatial and frequency domains through Spatial
Feature Extraction Block (SFEBlock) and Frequency Feature Extraction Block
(FFEBlock). We have further introduced an effective Adaptive Fusion Module
(AFM) for the dual-path feature aggregation. Extensive experiments on six
public deraining benchmarks and downstream vision tasks have demonstrated that
our proposed method not only outperforms the existing state-of-the-art
deraining method but also achieves visually pleasuring results with excellent
robustness on downstream vision tasks.
Related papers
- A Hybrid Transformer-Mamba Network for Single Image Deraining [70.64069487982916]
Existing deraining Transformers employ self-attention mechanisms with fixed-range windows or along channel dimensions.
We introduce a novel dual-branch hybrid Transformer-Mamba network, denoted as TransMamba, aimed at effectively capturing long-range rain-related dependencies.
arXiv Detail & Related papers (2024-08-31T10:03:19Z) - Adaptive Frequency Enhancement Network for Single Image Deraining [10.64622976628013]
We introduce a novel end-to-end Adaptive Frequency Enhancement Network (AFENet) specifically for single image deraining.
We employ convolutions of different scales to adaptively decompose image frequency bands, introduce a feature enhancement module, and present a novel interaction module.
This approach empowers the deraining network to eliminate diverse and complex rainy patterns and to reconstruct image details accurately.
arXiv Detail & Related papers (2024-07-19T13:24:05Z) - Gabor-guided transformer for single image deraining [2.330361251490783]
We propose a Gabor-guided tranformer (Gabformer) for single image deraining.
The focus on local texture features is enhanced by incorporating the information processed by the Gabor filter into the query vector.
Our method outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2024-03-12T07:41:51Z) - Multi-dimension Queried and Interacting Network for Stereo Image
Deraining [13.759978932686519]
We devise MQINet, which employs multi-dimension queries and interactions for stereo image deraining.
This module leverages dimension-wise queries that are independent of the input features.
We introduce an intra-view physics-aware attention (IPA) based on the inverse physical model of rainy images.
arXiv Detail & Related papers (2023-09-19T05:04:06Z) - Mutual Information-driven Triple Interaction Network for Efficient Image
Dehazing [54.168567276280505]
We propose a novel Mutual Information-driven Triple interaction Network (MITNet) for image dehazing.
The first stage, named amplitude-guided haze removal, aims to recover the amplitude spectrum of the hazy images for haze removal.
The second stage, named phase-guided structure refined, devotes to learning the transformation and refinement of the phase spectrum.
arXiv Detail & Related papers (2023-08-14T08:23:58Z) - Both Spatial and Frequency Cues Contribute to High-Fidelity Image
Inpainting [9.080472817672263]
Deep generative approaches have obtained great success in image inpainting recently.
Most generative inpainting networks suffer from either over-smooth results or aliasing artifacts.
We propose an effective Frequency-Spatial Complementary Network (FSCN) by exploiting rich semantic information in both spatial and frequency domains.
arXiv Detail & Related papers (2023-07-15T01:52:06Z) - DADFNet: Dual Attention and Dual Frequency-Guided Dehazing Network for
Video-Empowered Intelligent Transportation [79.18450119567315]
Adverse weather conditions pose severe challenges for video-based transportation surveillance.
We propose a dual attention and dual frequency-guided dehazing network (termed DADFNet) for real-time visibility enhancement.
arXiv Detail & Related papers (2023-04-19T11:55:30Z) - Spatial-Temporal Frequency Forgery Clue for Video Forgery Detection in
VIS and NIR Scenario [87.72258480670627]
Existing face forgery detection methods based on frequency domain find that the GAN forged images have obvious grid-like visual artifacts in the frequency spectrum compared to the real images.
This paper proposes a Cosine Transform-based Forgery Clue Augmentation Network (FCAN-DCT) to achieve a more comprehensive spatial-temporal feature representation.
arXiv Detail & Related papers (2022-07-05T09:27:53Z) - Adaptive Frequency Learning in Two-branch Face Forgery Detection [66.91715092251258]
We propose Adaptively learn Frequency information in the two-branch Detection framework, dubbed AFD.
We liberate our network from the fixed frequency transforms, and achieve better performance with our data- and task-dependent transform layers.
arXiv Detail & Related papers (2022-03-27T14:25:52Z) - EPMF: Efficient Perception-aware Multi-sensor Fusion for 3D Semantic Segmentation [62.210091681352914]
We study multi-sensor fusion for 3D semantic segmentation for many applications, such as autonomous driving and robotics.
In this work, we investigate a collaborative fusion scheme called perception-aware multi-sensor fusion (PMF)
We propose a two-stream network to extract features from the two modalities separately. The extracted features are fused by effective residual-based fusion modules.
arXiv Detail & Related papers (2021-06-21T10:47:26Z) - Deep Frequent Spatial Temporal Learning for Face Anti-Spoofing [9.435020319411311]
Face anti-spoofing is crucial for the security of face recognition system, by avoiding invaded with presentation attack.
Previous works have shown the effectiveness of using depth and temporal supervision for this task.
We propose a novel two stream FreqSaptialTemporalNet for face anti-spoofing which simultaneously takes advantage of frequent, spatial and temporal information.
arXiv Detail & Related papers (2020-01-20T06:02:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.