Improving underwater semantic segmentation with underwater image quality attention and muti-scale aggregation attention
- URL: http://arxiv.org/abs/2503.23422v1
- Date: Sun, 30 Mar 2025 12:47:56 GMT
- Title: Improving underwater semantic segmentation with underwater image quality attention and muti-scale aggregation attention
- Authors: Xin Zuo, Jiaran Jiang, Jifeng Shen, Wankou Yang,
- Abstract summary: UnderWater SegFormer (UWSegFormer) is a transformer-based framework for semantic segmentation of low-quality underwater images.<n>The proposed method has advantages in terms of segmentation completeness, boundary clarity, and subjective perceptual details when compared to SOTA methods.
- Score: 13.73105543582749
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Underwater image understanding is crucial for both submarine navigation and seabed exploration. However, the low illumination in underwater environments degrades the imaging quality, which in turn seriously deteriorates the performance of underwater semantic segmentation, particularly for outlining the object region boundaries. To tackle this issue, we present UnderWater SegFormer (UWSegFormer), a transformer-based framework for semantic segmentation of low-quality underwater images. Firstly, we propose the Underwater Image Quality Attention (UIQA) module. This module enhances the representation of highquality semantic information in underwater image feature channels through a channel self-attention mechanism. In order to address the issue of loss of imaging details due to the underwater environment, the Multi-scale Aggregation Attention(MAA) module is proposed. This module aggregates sets of semantic features at different scales by extracting discriminative information from high-level features,thus compensating for the semantic loss of detail in underwater objects. Finally, during training, we introduce Edge Learning Loss (ELL) in order to enhance the model's learning of underwater object edges and improve the model's prediction accuracy. Experiments conducted on the SUIM and DUT-USEG (DUT) datasets have demonstrated that the proposed method has advantages in terms of segmentation completeness, boundary clarity, and subjective perceptual details when compared to SOTA methods. In addition, the proposed method achieves the highest mIoU of 82.12 and 71.41 on the SUIM and DUT datasets, respectively. Code will be available at https://github.com/SAWRJJ/UWSegFormer.
Related papers
- Knowledge Distillation for Underwater Feature Extraction and Matching via GAN-synthesized Images [40.403791826344275]
In this paper, we aim to improve the robustness of the feature extraction and matching in the turbid underwater environment.
We first propose a novel adaptive GAN-synthesis method to estimate water parameters and underwater noise distribution.
We then introduce a general knowledge distillation framework compatible with different teacher models.
arXiv Detail & Related papers (2025-04-11T04:34:18Z) - FSSUWNet: Mitigating the Fragility of Pre-trained Models with Feature Enhancement for Few-Shot Semantic Segmentation in Underwater Images [4.19512807949895]
Few-Shot Semantic (FSS) has recently progressed in data-scarce domains.
We show that the existing FSS methods often struggle to generalize to underwater environments.
We propose FSSUWNet, a tailored FSS framework for underwater images with feature enhancement.
arXiv Detail & Related papers (2025-04-01T07:09:15Z) - A Unified Image-Dense Annotation Generation Model for Underwater Scenes [48.34534171882895]
This paper proposes a unified Text-to-Image and DEnse annotation generation method (TIDE) for underwater scenes.
It relies solely on text as input to simultaneously generate realistic underwater images and multiple highly consistent dense annotations.
We synthesize a large-scale underwater dataset using TIDE to validate the effectiveness of our method in underwater dense prediction tasks.
arXiv Detail & Related papers (2025-03-27T17:59:43Z) - WaterMono: Teacher-Guided Anomaly Masking and Enhancement Boosting for Robust Underwater Self-Supervised Monocular Depth Estimation [4.909989222186828]
We propose WaterMono, a novel framework for depth estimation and image enhancement.
It incorporates the following key measures: (1) We present a Teacher-Guided Anomaly Mask to identify dynamic regions within the images; (2) We employ depth information combined with the Underwater Image Formation Model to generate enhanced images, which in turn contribute to the depth estimation task; and (3) We utilize a rotated distillation strategy to enhance the model's rotational robustness.
arXiv Detail & Related papers (2024-06-19T08:49:45Z) - Diving into Underwater: Segment Anything Model Guided Underwater Salient Instance Segmentation and A Large-scale Dataset [60.14089302022989]
Underwater vision tasks often suffer from low segmentation accuracy due to the complex underwater circumstances.
We construct the first large-scale underwater salient instance segmentation dataset (USIS10K)
We propose an Underwater Salient Instance architecture based on Segment Anything Model (USIS-SAM) specifically for the underwater domain.
arXiv Detail & Related papers (2024-06-10T06:17:33Z) - Learning Heavily-Degraded Prior for Underwater Object Detection [59.5084433933765]
This paper seeks transferable prior knowledge from detector-friendly images.
It is based on statistical observations that, the heavily degraded regions of detector-friendly (DFUI) and underwater images have evident feature distribution gaps.
Our method with higher speeds and less parameters still performs better than transformer-based detectors.
arXiv Detail & Related papers (2023-08-24T12:32:46Z) - Synergistic Multiscale Detail Refinement via Intrinsic Supervision for
Underwater Image Enhancement [39.208417033777415]
We present intrinsic supervision (SMDR-IS) for enhancing underwater scene details, which contain multi-stages.
The ASISF module can precisely control and guide feature transmission across multi-degradation stages.
Bifocal Intrinsic-Context Attention Module (BICA) efficiently exploits multi-scale scene information in images.
arXiv Detail & Related papers (2023-08-23T05:40:55Z) - PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with
Dual-Discriminators [120.06891448820447]
How to obtain clear and visually pleasant images has become a common concern of people.
The task of underwater image enhancement (UIE) has also emerged as the times require.
In this paper, we propose a physical model-guided GAN model for UIE, referred to as PUGAN.
Our PUGAN outperforms state-of-the-art methods in both qualitative and quantitative metrics.
arXiv Detail & Related papers (2023-06-15T07:41:12Z) - DeepAqua: Self-Supervised Semantic Segmentation of Wetland Surface Water
Extent with SAR Images using Knowledge Distillation [44.99833362998488]
We present DeepAqua, a self-supervised deep learning model that eliminates the need for manual annotations during the training phase.
We exploit cases where optical- and radar-based water masks coincide, enabling the detection of both open and vegetated water surfaces.
Experimental results show that DeepAqua outperforms other unsupervised methods by improving accuracy by 7%, Intersection Over Union by 27%, and F1 score by 14%.
arXiv Detail & Related papers (2023-05-02T18:06:21Z) - Unpaired Overwater Image Defogging Using Prior Map Guided CycleGAN [60.257791714663725]
We propose a Prior map Guided CycleGAN (PG-CycleGAN) for defogging of images with overwater scenes.
The proposed method outperforms the state-of-the-art supervised, semi-supervised, and unsupervised defogging approaches.
arXiv Detail & Related papers (2022-12-23T03:00:28Z) - Semantic-aware Texture-Structure Feature Collaboration for Underwater
Image Enhancement [58.075720488942125]
Underwater image enhancement has become an attractive topic as a significant technology in marine engineering and aquatic robotics.
We develop an efficient and compact enhancement network in collaboration with a high-level semantic-aware pretrained model.
We also apply the proposed algorithm to the underwater salient object detection task to reveal the favorable semantic-aware ability for high-level vision tasks.
arXiv Detail & Related papers (2022-11-19T07:50:34Z) - A Benchmark dataset for both underwater image enhancement and underwater
object detection [34.25890702670983]
We provide a large-scale underwater object detection dataset with both bounding box annotations and high quality reference images.
The OUC dataset provides a platform to comprehensive study the influence of underwater image enhancement algorithms on the underwater object detection task.
arXiv Detail & Related papers (2020-06-29T03:12:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.