Multi-Focus Image Fusion based on Gradient Transform
- URL: http://arxiv.org/abs/2204.09777v1
- Date: Wed, 20 Apr 2022 20:35:12 GMT
- Title: Multi-Focus Image Fusion based on Gradient Transform
- Authors: Sultan Sevgi Turgut, Mustafa Oral
- Abstract summary: We introduce a novel gradient information-based multi-focus image fusion method that is robust for the aforementioned problems.
The proposed method is compared with 17 different novel and conventional techniques both visually and objectively.
It is observed that the proposed method is promising according to visual evaluation and 83.3% success is achieved by being first in five out of six metrics according to objective evaluation.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-focus image fusion is a challenging field of study that aims to provide
a completely focused image by integrating focused and un-focused pixels. Most
existing methods suffer from shift variance, misregistered images, and
data-dependent. In this study, we introduce a novel gradient information-based
multi-focus image fusion method that is robust for the aforementioned problems.
The proposed method first generates gradient images from original images by
using Halftoning-Inverse Halftoning (H-IH) transform. Then, Energy of Gradient
(EOG) and Standard Deviation functions are used as the focus measurement on the
gradient images to form a fused image. Finally, in order to enhance the fused
image a decision fusion approach is applied with the majority voting method.
The proposed method is compared with 17 different novel and conventional
techniques both visually and objectively. For objective evaluation, 6 different
quantitative metrics are used. It is observed that the proposed method is
promising according to visual evaluation and 83.3% success is achieved by being
first in five out of six metrics according to objective evaluation.
Related papers
- From Text to Pixels: A Context-Aware Semantic Synergy Solution for
Infrared and Visible Image Fusion [66.33467192279514]
We introduce a text-guided multi-modality image fusion method that leverages the high-level semantics from textual descriptions to integrate semantics from infrared and visible images.
Our method not only produces visually superior fusion results but also achieves a higher detection mAP over existing methods, achieving state-of-the-art results.
arXiv Detail & Related papers (2023-12-31T08:13:47Z) - Generation and Recombination for Multifocus Image Fusion with Free
Number of Inputs [17.32596568119519]
Multifocus image fusion is an effective way to overcome the limitation of optical lenses.
Previous methods assume that the focused areas of the two source images are complementary, making it impossible to achieve simultaneous fusion of multiple images.
In GRFusion, focus property detection of each source image can be implemented independently, enabling simultaneous fusion of multiple source images.
arXiv Detail & Related papers (2023-09-09T01:47:56Z) - A Task-guided, Implicitly-searched and Meta-initialized Deep Model for
Image Fusion [69.10255211811007]
We present a Task-guided, Implicit-searched and Meta- generalizationd (TIM) deep model to address the image fusion problem in a challenging real-world scenario.
Specifically, we propose a constrained strategy to incorporate information from downstream tasks to guide the unsupervised learning process of image fusion.
Within this framework, we then design an implicit search scheme to automatically discover compact architectures for our fusion model with high efficiency.
arXiv Detail & Related papers (2023-05-25T08:54:08Z) - Equivariant Multi-Modality Image Fusion [124.11300001864579]
We propose the Equivariant Multi-Modality imAge fusion paradigm for end-to-end self-supervised learning.
Our approach is rooted in the prior knowledge that natural imaging responses are equivariant to certain transformations.
Experiments confirm that EMMA yields high-quality fusion results for infrared-visible and medical images.
arXiv Detail & Related papers (2023-05-19T05:50:24Z) - DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion [144.9653045465908]
We propose a novel fusion algorithm based on the denoising diffusion probabilistic model (DDPM)
Our approach yields promising fusion results in infrared-visible image fusion and medical image fusion.
arXiv Detail & Related papers (2023-03-13T04:06:42Z) - Multi-focus thermal image fusion [0.34998703934432673]
The algorithm is based on local activity analysis and advanced pre-selection of images into fusion process.
The proposed algorithm is evaluated by half total error rate, root mean squared error, cross correlation and visual inspection.
arXiv Detail & Related papers (2022-03-16T10:27:33Z) - Unsupervised Image Fusion Using Deep Image Priors [7.549952136964352]
Deep Image Prior (DIP) method made it possible to do image restoration totally training-data-free.
This paper introduces a novel loss calculation structure, in the framework of DIP, while formulating image fusion as an inverse problem.
arXiv Detail & Related papers (2021-10-18T17:38:35Z) - Deep Reparametrization of Multi-Frame Super-Resolution and Denoising [167.42453826365434]
We propose a deep reparametrization of the maximum a posteriori formulation commonly employed in multi-frame image restoration tasks.
Our approach is derived by introducing a learned error metric and a latent representation of the target image.
We validate our approach through comprehensive experiments on burst denoising and burst super-resolution datasets.
arXiv Detail & Related papers (2021-08-18T17:57:02Z) - Towards Reducing Severe Defocus Spread Effects for Multi-Focus Image
Fusion via an Optimization Based Strategy [22.29205225281694]
Multi-focus image fusion (MFF) is a popular technique to generate an all-in-focus image.
This paper presents an optimization-based approach to reduce defocus spread effects.
Experiments conducted on the real-world dataset verify superiority of the proposed model.
arXiv Detail & Related papers (2020-12-29T09:26:41Z) - End-to-End Learning for Simultaneously Generating Decision Map and
Multi-Focus Image Fusion Result [7.564462759345851]
The aim of multi-focus image fusion is to gather focused regions of different images to generate a unique all-in-focus fused image.
Most of the existing deep learning structures failed to balance fusion quality and end-to-end implementation convenience.
We propose a cascade network to simultaneously generate decision map and fused result with an end-to-end training procedure.
arXiv Detail & Related papers (2020-10-17T09:09:51Z) - Depth image denoising using nuclear norm and learning graph model [107.51199787840066]
Group-based image restoration methods are more effective in gathering the similarity among patches.
For each patch, we find and group the most similar patches within a searching window.
The proposed method is superior to other current state-of-the-art denoising methods in both subjective and objective criterion.
arXiv Detail & Related papers (2020-08-09T15:12:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.