An Attention-Guided and Wavelet-Constrained Generative Adversarial
Network for Infrared and Visible Image Fusion
- URL: http://arxiv.org/abs/2210.11018v2
- Date: Mon, 24 Oct 2022 07:00:00 GMT
- Title: An Attention-Guided and Wavelet-Constrained Generative Adversarial
Network for Infrared and Visible Image Fusion
- Authors: Xiaowen Liu, Renhua Wang, Hongtao Huo, Xin Yang, Jing Li
- Abstract summary: We propose an attention-guided and wavelet-constrained GAN for infrared and visible image fusion (AWFGAN)
Specifically, we introduce the spatial attention modules (SAM) into the generator to obtain the spatial attention maps.
We extend the discrimination range of visible information to the wavelet subspace, which can force the generator to restore the high-frequency details of visible images.
- Score: 10.900528467160816
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The GAN-based infrared and visible image fusion methods have gained
ever-increasing attention due to its effectiveness and superiority. However,
the existing methods adopt the global pixel distribution of source images as
the basis for discrimination, which fails to focus on the key modality
information. Moreover, the dual-discriminator based methods suffer from the
confrontation between the discriminators. To this end, we propose an
attention-guided and wavelet-constrained GAN for infrared and visible image
fusion (AWFGAN). In this method, two unique discrimination strategies are
designed to improve the fusion performance. Specifically, we introduce the
spatial attention modules (SAM) into the generator to obtain the spatial
attention maps, and then the attention maps are utilized to force the
discrimination of infrared images to focus on the target regions. In addition,
we extend the discrimination range of visible information to the wavelet
subspace, which can force the generator to restore the high-frequency details
of visible images. Ablation experiments demonstrate the effectiveness of our
method in eliminating the confrontation between discriminators. And the
comparison experiments on public datasets demonstrate the effectiveness and
superiority of the proposed method.
Related papers
- GAN-HA: A generative adversarial network with a novel heterogeneous dual-discriminator network and a new attention-based fusion strategy for infrared and visible image fusion [0.1160897408844138]
Infrared and visible image fusion (IVIF) aims to preserve thermal radiation information from infrared images while integrating texture details from visible images.
Existing dual-discriminator generative adversarial networks (GANs) often rely on two structurally identical discriminators for learning.
This paper proposes a novel GAN with a heterogeneous dual-discriminator network and an attention-based fusion strategy.
arXiv Detail & Related papers (2024-04-24T17:06:52Z) - Decomposition-based and Interference Perception for Infrared and Visible
Image Fusion in Complex Scenes [4.919706769234434]
We propose a decomposition-based and interference perception image fusion method.
We classify the pixels of visible image from the degree of scattering of light transmission, based on which we then separate the detail and energy information of the image.
This refined decomposition facilitates the proposed model in identifying more interfering pixels that are in complex scenes.
arXiv Detail & Related papers (2024-02-03T09:27:33Z) - Cross-Modality Perturbation Synergy Attack for Person Re-identification [66.48494594909123]
The main challenge in cross-modality ReID lies in effectively dealing with visual differences between different modalities.
Existing attack methods have primarily focused on the characteristics of the visible image modality.
This study proposes a universal perturbation attack specifically designed for cross-modality ReID.
arXiv Detail & Related papers (2024-01-18T15:56:23Z) - Frequency Domain Nuances Mining for Visible-Infrared Person
Re-identification [75.87443138635432]
Existing methods mainly exploit the spatial information while ignoring the discriminative frequency information.
We propose a novel Frequency Domain Nuances Mining (FDNM) method to explore the cross-modality frequency domain information.
Our method outperforms the second-best method by 5.2% in Rank-1 accuracy and 5.8% in mAP on the SYSU-MM01 dataset.
arXiv Detail & Related papers (2024-01-04T09:19:54Z) - CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature
Ensemble for Multi-modality Image Fusion [72.8898811120795]
We propose a coupled contrastive learning network, dubbed CoCoNet, to realize infrared and visible image fusion.
Our method achieves state-of-the-art (SOTA) performance under both subjective and objective evaluation.
arXiv Detail & Related papers (2022-11-20T12:02:07Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - Infrared and Visible Image Fusion via Interactive Compensatory Attention
Adversarial Learning [7.995162257955025]
We propose a novel end-to-end mode based on generative adversarial training to achieve better fusion balance.
In particular, in the generator, we construct a multi-level encoder-decoder network with a triple path, and adopt infrared and visible paths to provide additional intensity and information gradient.
In addition, dual discriminators are designed to identify the similar distribution between fused result and source images, and the generator is optimized to produce a more balanced result.
arXiv Detail & Related papers (2022-03-29T08:28:14Z) - Exploring Separable Attention for Multi-Contrast MR Image
Super-Resolution [88.16655157395785]
We propose a separable attention network (comprising a priority attention and background separation attention) named SANet.
It can explore the foreground and background areas in the forward and reverse directions with the help of the auxiliary contrast.
It is the first model to explore a separable attention mechanism that uses the auxiliary contrast to predict the foreground and background regions.
arXiv Detail & Related papers (2021-09-03T05:53:07Z) - Interpreting Galaxy Deblender GAN from the Discriminator's Perspective [50.12901802952574]
This research focuses on behaviors of one of the network's major components, the Discriminator, which plays a vital role but is often overlooked.
We demonstrate that our method clearly reveals attention areas of the Discriminator when differentiating generated galaxy images from ground truth images.
arXiv Detail & Related papers (2020-01-17T04:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.