LRRNet: A Novel Representation Learning Guided Fusion Network for
Infrared and Visible Images
- URL: http://arxiv.org/abs/2304.05172v2
- Date: Sun, 16 Apr 2023 12:42:14 GMT
- Title: LRRNet: A Novel Representation Learning Guided Fusion Network for
Infrared and Visible Images
- Authors: Hui Li, Tianyang Xu, Xiao-Jun Wu, Jiwen Lu, Josef Kittler
- Abstract summary: We formulate the fusion task mathematically, and establish a connection between its optimal solution and the network architecture that can implement it.
In particular we adopt a learnable representation approach to the fusion task, in which the construction of the fusion network architecture is guided by the optimisation algorithm producing the learnable model.
Based on this novel network architecture, an end-to-end lightweight fusion network is constructed to fuse infrared and visible light images.
- Score: 98.36300655482196
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning based fusion methods have been achieving promising performance
in image fusion tasks. This is attributed to the network architecture that
plays a very important role in the fusion process. However, in general, it is
hard to specify a good fusion architecture, and consequently, the design of
fusion networks is still a black art, rather than science. To address this
problem, we formulate the fusion task mathematically, and establish a
connection between its optimal solution and the network architecture that can
implement it. This approach leads to a novel method proposed in the paper of
constructing a lightweight fusion network. It avoids the time-consuming
empirical network design by a trial-and-test strategy. In particular we adopt a
learnable representation approach to the fusion task, in which the construction
of the fusion network architecture is guided by the optimisation algorithm
producing the learnable model. The low-rank representation (LRR) objective is
the foundation of our learnable model. The matrix multiplications, which are at
the heart of the solution are transformed into convolutional operations, and
the iterative process of optimisation is replaced by a special feed-forward
network. Based on this novel network architecture, an end-to-end lightweight
fusion network is constructed to fuse infrared and visible light images. Its
successful training is facilitated by a detail-to-semantic information loss
function proposed to preserve the image details and to enhance the salient
features of the source images. Our experiments show that the proposed fusion
network exhibits better fusion performance than the state-of-the-art fusion
methods on public datasets. Interestingly, our network requires a fewer
training parameters than other existing methods. The codes are available at
https://github.com/hli1221/imagefusion-LRRNet
Related papers
- A Task-guided, Implicitly-searched and Meta-initialized Deep Model for
Image Fusion [69.10255211811007]
We present a Task-guided, Implicit-searched and Meta- generalizationd (TIM) deep model to address the image fusion problem in a challenging real-world scenario.
Specifically, we propose a constrained strategy to incorporate information from downstream tasks to guide the unsupervised learning process of image fusion.
Within this framework, we then design an implicit search scheme to automatically discover compact architectures for our fusion model with high efficiency.
arXiv Detail & Related papers (2023-05-25T08:54:08Z) - Searching a Compact Architecture for Robust Multi-Exposure Image Fusion [55.37210629454589]
Two major stumbling blocks hinder the development, including pixel misalignment and inefficient inference.
This study introduces an architecture search-based paradigm incorporating self-alignment and detail repletion modules for robust multi-exposure image fusion.
The proposed method outperforms various competitive schemes, achieving a noteworthy 3.19% improvement in PSNR for general scenarios and an impressive 23.5% enhancement in misaligned scenarios.
arXiv Detail & Related papers (2023-05-20T17:01:52Z) - CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature
Ensemble for Multi-modality Image Fusion [72.8898811120795]
We propose a coupled contrastive learning network, dubbed CoCoNet, to realize infrared and visible image fusion.
Our method achieves state-of-the-art (SOTA) performance under both subjective and objective evaluation.
arXiv Detail & Related papers (2022-11-20T12:02:07Z) - Unsupervised Image Fusion Method based on Feature Mutual Mapping [16.64607158983448]
We propose an unsupervised adaptive image fusion method to address the above issues.
We construct a global map to measure the connections of pixels between the input source images.
Our method achieves superior performance in both visual perception and objective evaluation.
arXiv Detail & Related papers (2022-01-25T07:50:14Z) - TransFuse: A Unified Transformer-based Image Fusion Framework using
Self-supervised Learning [5.849513679510834]
Image fusion is a technique to integrate information from multiple source images with complementary information to improve the richness of a single image.
Two-stage methods avoid the need of large amount of task-specific training data by training encoder-decoder network on large natural image datasets.
We propose a destruction-reconstruction based self-supervised training scheme to encourage the network to learn task-specific features.
arXiv Detail & Related papers (2022-01-19T07:30:44Z) - Image Fusion Transformer [75.71025138448287]
In image fusion, images obtained from different sensors are fused to generate a single image with enhanced information.
In recent years, state-of-the-art methods have adopted Convolution Neural Networks (CNNs) to encode meaningful features for image fusion.
We propose a novel Image Fusion Transformer (IFT) where we develop a transformer-based multi-scale fusion strategy.
arXiv Detail & Related papers (2021-07-19T16:42:49Z) - RFN-Nest: An end-to-end residual fusion network for infrared and visible
images [37.935940961760785]
We propose an end-to-end fusion network architecture (RFN-Nest) for infrared and visible image fusion.
A novel detail-preserving loss function, and a feature enhancing loss function are proposed to train RFN.
The experimental results on public domain data sets show that, compared with the existing methods, our end-to-end fusion network delivers a better performance than the state-of-the-art methods.
arXiv Detail & Related papers (2021-03-07T07:29:50Z) - NestFuse: An Infrared and Visible Image Fusion Architecture based on
Nest Connection and Spatial/Channel Attention Models [12.16870022547833]
We propose a novel method for infrared and visible image fusion.
We develop nest connection-based network and spatial/channel attention models.
Experiments are performed on publicly available datasets.
arXiv Detail & Related papers (2020-07-01T08:46:23Z) - RoutedFusion: Learning Real-time Depth Map Fusion [73.0378509030908]
We present a novel real-time capable machine learning-based method for depth map fusion.
We propose a neural network that predicts non-linear updates to better account for typical fusion errors.
Our network is composed of a 2D depth routing network and a 3D depth fusion network which efficiently handle sensor-specific noise and outliers.
arXiv Detail & Related papers (2020-01-13T16:46:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.