MuLA-GAN: Multi-Level Attention GAN for Enhanced Underwater Visibility
- URL: http://arxiv.org/abs/2312.15633v1
- Date: Mon, 25 Dec 2023 07:33:47 GMT
- Title: MuLA-GAN: Multi-Level Attention GAN for Enhanced Underwater Visibility
- Authors: Ahsan Baidar Bakht, Zikai Jia, Muhayy ud Din, Waseem Akram, Lyes Saad
Soud, Lakmal Seneviratne, Defu Lin, Shaoming He and Irfan Hussain
- Abstract summary: We introduce MuLA-GAN, a novel approach that leverages the synergistic power of Geneversarative Adrial Networks (GANs) and Multi-Level Attention mechanisms for comprehensive underwater image enhancement.
Our model excels in capturing and preserving intricate details in underwater imagery, essential for various applications.
This work not only addresses a significant research gap in underwater image enhancement but also underscores the pivotal role of Multi-Level Attention in enhancing GANs.
- Score: 1.9272863690919875
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The underwater environment presents unique challenges, including color
distortions, reduced contrast, and blurriness, hindering accurate analysis. In
this work, we introduce MuLA-GAN, a novel approach that leverages the
synergistic power of Generative Adversarial Networks (GANs) and Multi-Level
Attention mechanisms for comprehensive underwater image enhancement. The
integration of Multi-Level Attention within the GAN architecture significantly
enhances the model's capacity to learn discriminative features crucial for
precise image restoration. By selectively focusing on relevant spatial and
multi-level features, our model excels in capturing and preserving intricate
details in underwater imagery, essential for various applications. Extensive
qualitative and quantitative analyses on diverse datasets, including UIEB test
dataset, UIEB challenge dataset, U45, and UCCS dataset, highlight the superior
performance of MuLA-GAN compared to existing state-of-the-art methods.
Experimental evaluations on a specialized dataset tailored for bio-fouling and
aquaculture applications demonstrate the model's robustness in challenging
environmental conditions. On the UIEB test dataset, MuLA-GAN achieves
exceptional PSNR (25.59) and SSIM (0.893) scores, surpassing Water-Net, the
second-best model, with scores of 24.36 and 0.885, respectively. This work not
only addresses a significant research gap in underwater image enhancement but
also underscores the pivotal role of Multi-Level Attention in enhancing GANs,
providing a novel and comprehensive framework for restoring underwater image
quality.
Related papers
- UW-SDF: Exploiting Hybrid Geometric Priors for Neural SDF Reconstruction from Underwater Multi-view Monocular Images [63.32490897641344]
We propose a framework for reconstructing target objects from multi-view underwater images based on neural SDF.
We introduce hybrid geometric priors to optimize the reconstruction process, markedly enhancing the quality and efficiency of neural SDF reconstruction.
arXiv Detail & Related papers (2024-10-10T16:33:56Z) - UIE-UnFold: Deep Unfolding Network with Color Priors and Vision Transformer for Underwater Image Enhancement [27.535028176427623]
Underwater image enhancement (UIE) plays a crucial role in various marine applications.
Current learning-based approaches frequently lack explicit prior knowledge about the physical processes involved in underwater image formation.
This paper proposes a novel deep unfolding network (DUN) for UIE that integrates color priors and inter-stage feature incorporation.
arXiv Detail & Related papers (2024-08-20T08:48:33Z) - Opinion-Unaware Blind Image Quality Assessment using Multi-Scale Deep Feature Statistics [54.08757792080732]
We propose integrating deep features from pre-trained visual models with a statistical analysis model to achieve opinion-unaware BIQA (OU-BIQA)
Our proposed model exhibits superior consistency with human visual perception compared to state-of-the-art BIQA models.
arXiv Detail & Related papers (2024-05-29T06:09:34Z) - Separated Attention: An Improved Cycle GAN Based Under Water Image Enhancement Method [0.0]
We have utilized the cycle consistent learning technique of the state-of-the-art Cycle GAN model with modification in the loss function.
We trained the Cycle GAN model with the modified loss functions on the benchmarked Enhancing Underwater Visual Perception dataset.
The upgraded images provide better results from conventional models and further for under water navigation, pose estimation, saliency prediction, object detection and tracking.
arXiv Detail & Related papers (2024-04-11T11:12:06Z) - Diffusion Model Based Visual Compensation Guidance and Visual Difference
Analysis for No-Reference Image Quality Assessment [82.13830107682232]
We propose a novel class of state-of-the-art (SOTA) generative model, which exhibits the capability to model intricate relationships.
We devise a new diffusion restoration network that leverages the produced enhanced image and noise-containing images.
Two visual evaluation branches are designed to comprehensively analyze the obtained high-level feature information.
arXiv Detail & Related papers (2024-02-22T09:39:46Z) - DGNet: Dynamic Gradient-Guided Network for Water-Related Optics Image
Enhancement [77.0360085530701]
Underwater image enhancement (UIE) is a challenging task due to the complex degradation caused by underwater environments.
Previous methods often idealize the degradation process, and neglect the impact of medium noise and object motion on the distribution of image features.
Our approach utilizes predicted images to dynamically update pseudo-labels, adding a dynamic gradient to optimize the network's gradient space.
arXiv Detail & Related papers (2023-12-12T06:07:21Z) - PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with
Dual-Discriminators [120.06891448820447]
How to obtain clear and visually pleasant images has become a common concern of people.
The task of underwater image enhancement (UIE) has also emerged as the times require.
In this paper, we propose a physical model-guided GAN model for UIE, referred to as PUGAN.
Our PUGAN outperforms state-of-the-art methods in both qualitative and quantitative metrics.
arXiv Detail & Related papers (2023-06-15T07:41:12Z) - Semantic-aware Texture-Structure Feature Collaboration for Underwater
Image Enhancement [58.075720488942125]
Underwater image enhancement has become an attractive topic as a significant technology in marine engineering and aquatic robotics.
We develop an efficient and compact enhancement network in collaboration with a high-level semantic-aware pretrained model.
We also apply the proposed algorithm to the underwater salient object detection task to reveal the favorable semantic-aware ability for high-level vision tasks.
arXiv Detail & Related papers (2022-11-19T07:50:34Z) - A Novel Underwater Image Enhancement and Improved Underwater Biological
Detection Pipeline [8.326477369707122]
This paper proposes a novel method for capturing feature information, which adds the convolutional block attention module (CBAM) to the YOLOv5 backbone.
The interference of underwater creature characteristics on object characteristics is decreased, and the output of the backbone network to object information is enhanced.
arXiv Detail & Related papers (2022-05-20T14:18:17Z) - UIF: An Objective Quality Assessment for Underwater Image Enhancement [17.145844358253164]
We propose an Underwater Image Fidelity (UIF) metric for objective evaluation of enhanced underwater images.
By exploiting the statistical features of these images, we present to extract naturalness-related, sharpness-related, and structure-related features.
Experimental results confirm that the proposed UIF outperforms a variety of underwater and general-purpose image quality metrics.
arXiv Detail & Related papers (2022-05-19T08:43:47Z) - Simultaneous Enhancement and Super-Resolution of Underwater Imagery for
Improved Visual Perception [17.403133838762447]
We introduce and tackle the simultaneous enhancement and super-resolution (SESR) problem for underwater robot vision.
We present Deep SESR, a residual-in-residual network-based generative model that can learn to restore perceptual image qualities at 2x, 3x, or 4x higher spatial resolution.
arXiv Detail & Related papers (2020-02-04T07:07:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.