MFIF-GAN: A New Generative Adversarial Network for Multi-Focus Image
Fusion
- URL: http://arxiv.org/abs/2009.09718v4
- Date: Mon, 9 Nov 2020 03:36:53 GMT
- Title: MFIF-GAN: A New Generative Adversarial Network for Multi-Focus Image
Fusion
- Authors: Yicheng Wang, Shuang Xu, Junmin Liu, Zixiang Zhao, Chunxia Zhang,
Jiangshe Zhang
- Abstract summary: Multi-Focus Image Fusion (MFIF) is a promising technique to obtain all-in-focus images.
One of the research trends of MFIF is to avoid the defocus spread effect (DSE) around the focus/defocus boundary (FDB)
We propose a network termed MFIF-GAN to generate focus maps in which the foreground region are correctly larger than the corresponding objects.
- Score: 29.405149234582623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-Focus Image Fusion (MFIF) is a promising image enhancement technique to
obtain all-in-focus images meeting visual needs and it is a precondition of
other computer vision tasks. One of the research trends of MFIF is to avoid the
defocus spread effect (DSE) around the focus/defocus boundary (FDB). In this
paper,we propose a network termed MFIF-GAN to attenuate the DSE by generating
focus maps in which the foreground region are correctly larger than the
corresponding objects. The Squeeze and Excitation Residual module is employed
in the network. By combining the prior knowledge of training condition, this
network is trained on a synthetic dataset based on an {\alpha}-matte model. In
addition, the reconstruction and gradient regularization terms are combined in
the loss functions to enhance the boundary details and improve the quality of
fused images. Extensive experiments demonstrate that the MFIF-GAN outperforms
several state-of-the-art (SOTA) methods in visual perception, quantitative
analysis as well as efficiency. Moreover, the edge diffusion and contraction
module is firstly proposed to verify that focus maps generated by our method
are accurate at the pixel level.
Related papers
- MFCLIP: Multi-modal Fine-grained CLIP for Generalizable Diffusion Face Forgery Detection [64.29452783056253]
The rapid development of photo-realistic face generation methods has raised significant concerns in society and academia.
Although existing approaches mainly capture face forgery patterns using image modality, other modalities like fine-grained noises and texts are not fully explored.
We propose a novel multi-modal fine-grained CLIP (MFCLIP) model, which mines comprehensive and fine-grained forgery traces across image-noise modalities.
arXiv Detail & Related papers (2024-09-15T13:08:59Z) - Fiducial Focus Augmentation for Facial Landmark Detection [4.433764381081446]
We propose a novel image augmentation technique to enhance the model's understanding of facial structures.
We employ a Siamese architecture-based training mechanism with a Deep Canonical Correlation Analysis (DCCA)-based loss.
Our approach outperforms multiple state-of-the-art approaches across various benchmark datasets.
arXiv Detail & Related papers (2024-02-23T01:34:00Z) - JoReS-Diff: Joint Retinex and Semantic Priors in Diffusion Model for Low-light Image Enhancement [69.6035373784027]
Low-light image enhancement (LLIE) has achieved promising performance by employing conditional diffusion models.
Previous methods may neglect the importance of a sufficient formulation of task-specific condition strategy.
We propose JoReS-Diff, a novel approach that incorporates Retinex- and semantic-based priors as the additional pre-processing condition.
arXiv Detail & Related papers (2023-12-20T08:05:57Z) - Guided Image Restoration via Simultaneous Feature and Image Guided
Fusion [67.30078778732998]
We propose a Simultaneous Feature and Image Guided Fusion (SFIGF) network.
It considers feature and image-level guided fusion following the guided filter (GF) mechanism.
Since guided fusion is implemented in both feature and image domains, the proposed SFIGF is expected to faithfully reconstruct both contextual and textual information.
arXiv Detail & Related papers (2023-12-14T12:15:45Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - Bridging the Gap between Multi-focus and Multi-modal: A Focused
Integration Framework for Multi-modal Image Fusion [5.417493475406649]
Multi-modal image fusion (MMIF) integrates valuable information from different modality images into a fused one.
This paper proposes a MMIF framework for joint focused integration and modalities information extraction.
The proposed algorithm can surpass the state-of-the-art methods in visual perception and quantitative evaluation.
arXiv Detail & Related papers (2023-11-03T12:58:39Z) - BigFUSE: Global Context-Aware Image Fusion in Dual-View Light-Sheet
Fluorescence Microscopy with Image Formation Prior [40.22867974147714]
We propose BigFUSE, a global context-aware image fuser that stabilizes image fusion in light-sheet fluorescence microscopy (LSFM)
Inspired by the image formation prior in dual-view LSFM, image fusion is considered as estimating a focus-defocus boundary using Bayes Theorem.
Competitive experimental results show that BigFUSE is the first dual-view LSFM fuser that is able to exclude structured artifacts when fusing information.
arXiv Detail & Related papers (2023-09-05T00:05:09Z) - Hybrid-Supervised Dual-Search: Leveraging Automatic Learning for
Loss-free Multi-Exposure Image Fusion [60.221404321514086]
Multi-exposure image fusion (MEF) has emerged as a prominent solution to address the limitations of digital imaging in representing varied exposure levels.
This paper presents a Hybrid-Supervised Dual-Search approach for MEF, dubbed HSDS-MEF, which introduces a bi-level optimization search scheme for automatic design of both network structures and loss functions.
arXiv Detail & Related papers (2023-09-03T08:07:26Z) - An Interactively Reinforced Paradigm for Joint Infrared-Visible Image
Fusion and Saliency Object Detection [59.02821429555375]
This research focuses on the discovery and localization of hidden objects in the wild and serves unmanned systems.
Through empirical analysis, infrared and visible image fusion (IVIF) enables hard-to-find objects apparent.
multimodal salient object detection (SOD) accurately delineates the precise spatial location of objects within the picture.
arXiv Detail & Related papers (2023-05-17T06:48:35Z) - Towards Reducing Severe Defocus Spread Effects for Multi-Focus Image
Fusion via an Optimization Based Strategy [22.29205225281694]
Multi-focus image fusion (MFF) is a popular technique to generate an all-in-focus image.
This paper presents an optimization-based approach to reduce defocus spread effects.
Experiments conducted on the real-world dataset verify superiority of the proposed model.
arXiv Detail & Related papers (2020-12-29T09:26:41Z) - FFusionCGAN: An end-to-end fusion method for few-focus images using
conditional GAN in cytopathological digital slides [0.0]
Multi-focus image fusion technologies compress different focus depth images into an image in which most objects are in focus.
This paper proposes a novel method for generating fused images from single-focus or few-focus images based on conditional generative adversarial network (GAN)
By integrating the network into the generative model, the quality of the generated fused images is effectively improved.
arXiv Detail & Related papers (2020-01-03T02:13:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.