Learning a Graph Neural Network with Cross Modality Interaction for
Image Fusion
- URL: http://arxiv.org/abs/2308.03256v1
- Date: Mon, 7 Aug 2023 02:25:06 GMT
- Title: Learning a Graph Neural Network with Cross Modality Interaction for
Image Fusion
- Authors: Jiawei Li, Jiansheng Chen, Jinyuan Liu, Huimin Ma
- Abstract summary: Infrared and visible image fusion has gradually proved to be a vital fork in the field of multi-modality imaging technologies.
We propose an interactive graph neural network (GNN)-based architecture between cross modality for fusion, called IGNet.
Our IGNet can generate visually appealing fused images while scoring averagely 2.59% mAP@.5 and 7.77% mIoU higher in detection and segmentation.
- Score: 23.296468921842948
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Infrared and visible image fusion has gradually proved to be a vital fork in
the field of multi-modality imaging technologies. In recent developments,
researchers not only focus on the quality of fused images but also evaluate
their performance in downstream tasks. Nevertheless, the majority of methods
seldom put their eyes on the mutual learning from different modalities,
resulting in fused images lacking significant details and textures. To overcome
this issue, we propose an interactive graph neural network (GNN)-based
architecture between cross modality for fusion, called IGNet. Specifically, we
first apply a multi-scale extractor to achieve shallow features, which are
employed as the necessary input to build graph structures. Then, the graph
interaction module can construct the extracted intermediate features of the
infrared/visible branch into graph structures. Meanwhile, the graph structures
of two branches interact for cross-modality and semantic learning, so that
fused images can maintain the important feature expressions and enhance the
performance of downstream tasks. Besides, the proposed leader nodes can improve
information propagation in the same modality. Finally, we merge all graph
features to get the fusion result. Extensive experiments on different datasets
(TNO, MFNet and M3FD) demonstrate that our IGNet can generate visually
appealing fused images while scoring averagely 2.59% mAP@.5 and 7.77% mIoU
higher in detection and segmentation than the compared state-of-the-art
methods. The source code of the proposed IGNet can be available at
https://github.com/lok-18/IGNet.
Related papers
- Mew: Multiplexed Immunofluorescence Image Analysis through an Efficient Multiplex Network [84.88767228835928]
We introduce Mew, a novel framework designed to efficiently process mIF images through the lens of multiplex network.
Mew innovatively constructs a multiplex network comprising two distinct layers: a Voronoi network for geometric information and a Cell-type network for capturing cell-wise homogeneity.
This framework equips a scalable and efficient Graph Neural Network (GNN), capable of processing the entire graph during training.
arXiv Detail & Related papers (2024-07-25T08:22:30Z) - A Semantic-Aware and Multi-Guided Network for Infrared-Visible Image Fusion [41.34335755315773]
Multi-modality image fusion aims at fusing specific-modality and shared-modality information from two source images.
We propose a three-branch encoder-decoder architecture along with corresponding fusion layers as the fusion strategy.
Our method has obtained competitive results compared with state-of-the-art methods in visible/infrared image fusion and medical image fusion tasks.
arXiv Detail & Related papers (2024-06-11T09:32:40Z) - From Text to Pixels: A Context-Aware Semantic Synergy Solution for
Infrared and Visible Image Fusion [66.33467192279514]
We introduce a text-guided multi-modality image fusion method that leverages the high-level semantics from textual descriptions to integrate semantics from infrared and visible images.
Our method not only produces visually superior fusion results but also achieves a higher detection mAP over existing methods, achieving state-of-the-art results.
arXiv Detail & Related papers (2023-12-31T08:13:47Z) - Graph Representation Learning for Infrared and Visible Image Fusion [19.756524363404534]
Infrared and visible image fusion aims to extract complementary features to synthesize a single fused image.
Many methods employ convolutional neural networks (CNNs) to extract local features.
CNNs fail to consider the image's non-local self-similarity (NLss)
arXiv Detail & Related papers (2023-11-01T04:46:20Z) - An Interactively Reinforced Paradigm for Joint Infrared-Visible Image
Fusion and Saliency Object Detection [59.02821429555375]
This research focuses on the discovery and localization of hidden objects in the wild and serves unmanned systems.
Through empirical analysis, infrared and visible image fusion (IVIF) enables hard-to-find objects apparent.
multimodal salient object detection (SOD) accurately delineates the precise spatial location of objects within the picture.
arXiv Detail & Related papers (2023-05-17T06:48:35Z) - CDDFuse: Correlation-Driven Dual-Branch Feature Decomposition for
Multi-Modality Image Fusion [138.40422469153145]
We propose a novel Correlation-Driven feature Decomposition Fusion (CDDFuse) network.
We show that CDDFuse achieves promising results in multiple fusion tasks, including infrared-visible image fusion and medical image fusion.
arXiv Detail & Related papers (2022-11-26T02:40:28Z) - CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature
Ensemble for Multi-modality Image Fusion [72.8898811120795]
We propose a coupled contrastive learning network, dubbed CoCoNet, to realize infrared and visible image fusion.
Our method achieves state-of-the-art (SOTA) performance under both subjective and objective evaluation.
arXiv Detail & Related papers (2022-11-20T12:02:07Z) - Learnable Graph Convolutional Network and Feature Fusion for Multi-view
Learning [30.74535386745822]
This paper proposes a joint deep learning framework called Learnable Graph Convolutional Network and Feature Fusion (LGCN-FF)
It consists of two stages: feature fusion network and learnable graph convolutional network.
The proposed LGCN-FF is validated to be superior to various state-of-the-art methods in multi-view semi-supervised classification.
arXiv Detail & Related papers (2022-11-16T19:07:12Z) - FuseVis: Interpreting neural networks for image fusion using per-pixel
saliency visualization [10.156766309614113]
Unsupervised learning based convolutional neural networks (CNNs) have been utilized for different types of image fusion tasks.
It is challenging to analyze the reliability of these CNNs for the image fusion tasks since no groundtruth is available.
We present a novel real-time visualization tool, named FuseVis, with which the end-user can compute per-pixel saliency maps.
arXiv Detail & Related papers (2020-12-06T10:03:02Z) - Multi-hop Attention Graph Neural Network [70.21119504298078]
Multi-hop Attention Graph Neural Network (MAGNA) is a principled way to incorporate multi-hop context information into every layer of attention computation.
We show that MAGNA captures large-scale structural information in every layer, and has a low-pass effect that eliminates noisy high-frequency information from graph data.
arXiv Detail & Related papers (2020-09-29T22:41:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.