Breaking Free from Fusion Rule: A Fully Semantic-driven Infrared and
Visible Image Fusion
- URL: http://arxiv.org/abs/2211.12286v1
- Date: Tue, 22 Nov 2022 13:59:59 GMT
- Title: Breaking Free from Fusion Rule: A Fully Semantic-driven Infrared and
Visible Image Fusion
- Authors: Yuhui Wu, Zhu Liu, Jinyuan Liu, Xin Fan, Risheng Liu
- Abstract summary: Infrared and visible image fusion plays a vital role in the field of computer vision.
Previous approaches make efforts to design various fusion rules in the loss functions.
We develop a semantic-level fusion network to sufficiently utilize the semantic guidance.
- Score: 51.22863068854784
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Infrared and visible image fusion plays a vital role in the field of computer
vision. Previous approaches make efforts to design various fusion rules in the
loss functions. However, these experimental designed fusion rules make the
methods more and more complex. Besides, most of them only focus on boosting the
visual effects, thus showing unsatisfactory performance for the follow-up
high-level vision tasks. To address these challenges, in this letter, we
develop a semantic-level fusion network to sufficiently utilize the semantic
guidance, emancipating the experimental designed fusion rules. In addition, to
achieve a better semantic understanding of the feature fusion process, a fusion
block based on the transformer is presented in a multi-scale manner. Moreover,
we devise a regularization loss function, together with a training strategy, to
fully use semantic guidance from the high-level vision tasks. Compared with
state-of-the-art methods, our method does not depend on the hand-crafted fusion
loss function. Still, it achieves superior performance on visual quality along
with the follow-up high-level vision tasks.
Related papers
- MaeFuse: Transferring Omni Features with Pretrained Masked Autoencoders for Infrared and Visible Image Fusion via Guided Training [57.18758272617101]
MaeFuse is a novel autoencoder model designed for infrared and visible image fusion (IVIF)
Our model utilizes a pretrained encoder from Masked Autoencoders (MAE), which facilities the omni features extraction for low-level reconstruction and high-level vision tasks.
MaeFuse not only introduces a novel perspective in the realm of fusion techniques but also stands out with impressive performance across various public datasets.
arXiv Detail & Related papers (2024-04-17T02:47:39Z) - From Text to Pixels: A Context-Aware Semantic Synergy Solution for
Infrared and Visible Image Fusion [66.33467192279514]
We introduce a text-guided multi-modality image fusion method that leverages the high-level semantics from textual descriptions to integrate semantics from infrared and visible images.
Our method not only produces visually superior fusion results but also achieves a higher detection mAP over existing methods, achieving state-of-the-art results.
arXiv Detail & Related papers (2023-12-31T08:13:47Z) - A Task-guided, Implicitly-searched and Meta-initialized Deep Model for
Image Fusion [69.10255211811007]
We present a Task-guided, Implicit-searched and Meta- generalizationd (TIM) deep model to address the image fusion problem in a challenging real-world scenario.
Specifically, we propose a constrained strategy to incorporate information from downstream tasks to guide the unsupervised learning process of image fusion.
Within this framework, we then design an implicit search scheme to automatically discover compact architectures for our fusion model with high efficiency.
arXiv Detail & Related papers (2023-05-25T08:54:08Z) - Bi-level Dynamic Learning for Jointly Multi-modality Image Fusion and
Beyond [50.556961575275345]
We build an image fusion module to fuse complementary characteristics and cascade dual task-related modules.
We develop an efficient first-order approximation to compute corresponding gradients and present dynamic weighted aggregation to balance the gradients for fusion learning.
arXiv Detail & Related papers (2023-05-11T10:55:34Z) - CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature
Ensemble for Multi-modality Image Fusion [72.8898811120795]
We propose a coupled contrastive learning network, dubbed CoCoNet, to realize infrared and visible image fusion.
Our method achieves state-of-the-art (SOTA) performance under both subjective and objective evaluation.
arXiv Detail & Related papers (2022-11-20T12:02:07Z) - Unsupervised Image Fusion Method based on Feature Mutual Mapping [16.64607158983448]
We propose an unsupervised adaptive image fusion method to address the above issues.
We construct a global map to measure the connections of pixels between the input source images.
Our method achieves superior performance in both visual perception and objective evaluation.
arXiv Detail & Related papers (2022-01-25T07:50:14Z) - TGFuse: An Infrared and Visible Image Fusion Approach Based on
Transformer and Generative Adversarial Network [15.541268697843037]
We propose an infrared and visible image fusion algorithm based on a lightweight transformer module and adversarial learning.
Inspired by the global interaction power, we use the transformer technique to learn the effective global fusion relations.
The experimental performance demonstrates the effectiveness of the proposed modules, with superior improvement against the state-of-the-art.
arXiv Detail & Related papers (2022-01-25T07:43:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.