Iterative, Deep Synthetic Aperture Sonar Image Segmentation
- URL: http://arxiv.org/abs/2203.15082v1
- Date: Mon, 28 Mar 2022 20:41:24 GMT
- Title: Iterative, Deep Synthetic Aperture Sonar Image Segmentation
- Authors: Yung-Chen Sun, Isaac D. Gerg, and Vishal Monga
- Abstract summary: We propose an unsupervised learning framework called Iterative Deep Unsupervised (IDUS) for SAS image segmentation.
IDUS can be divided into four main steps: 1) A deep network estimates class assignments; 2) Low-level image features from the deep network are clustered into superpixels; 3) Superpixels are clustered into class assignments; 4) Resulting pseudo-labels are used for loss backpropagation of the deep network prediction.
A comparison of IDUS to current state-of-the-art methods on a realistic benchmark dataset for SAS image segmentation demonstrates the benefits of our proposal.
- Score: 21.319490900396474
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Synthetic aperture sonar (SAS) systems produce high-resolution images of the
seabed environment. Moreover, deep learning has demonstrated superior ability
in finding robust features for automating imagery analysis. However, the
success of deep learning is conditioned on having lots of labeled training
data, but obtaining generous pixel-level annotations of SAS imagery is often
practically infeasible. This challenge has thus far limited the adoption of
deep learning methods for SAS segmentation. Algorithms exist to segment SAS
imagery in an unsupervised manner, but they lack the benefit of
state-of-the-art learning methods and the results present significant room for
improvement. In view of the above, we propose a new iterative algorithm for
unsupervised SAS image segmentation combining superpixel formation, deep
learning, and traditional clustering methods. We call our method Iterative Deep
Unsupervised Segmentation (IDUS). IDUS is an unsupervised learning framework
that can be divided into four main steps: 1) A deep network estimates class
assignments. 2) Low-level image features from the deep network are clustered
into superpixels. 3) Superpixels are clustered into class assignments (which we
call pseudo-labels) using $k$-means. 4) Resulting pseudo-labels are used for
loss backpropagation of the deep network prediction. These four steps are
performed iteratively until convergence. A comparison of IDUS to current
state-of-the-art methods on a realistic benchmark dataset for SAS image
segmentation demonstrates the benefits of our proposal even as the IDUS incurs
a much lower computational burden during inference (actual labeling of a test
image). Finally, we also develop a semi-supervised (SS) extension of IDUS
called IDSS and demonstrate experimentally that it can further enhance
performance while outperforming supervised alternatives that exploit the same
labeled training imagery.
Related papers
- Semi-supervised Semantic Segmentation for Remote Sensing Images via Multi-scale Uncertainty Consistency and Cross-Teacher-Student Attention [59.19580789952102]
This paper proposes a novel semi-supervised Multi-Scale Uncertainty and Cross-Teacher-Student Attention (MUCA) model for RS image semantic segmentation tasks.
MUCA constrains the consistency among feature maps at different layers of the network by introducing a multi-scale uncertainty consistency regularization.
MUCA utilizes a Cross-Teacher-Student attention mechanism to guide the student network, guiding the student network to construct more discriminative feature representations.
arXiv Detail & Related papers (2025-01-18T11:57:20Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - A Dual-branch Self-supervised Representation Learning Framework for
Tumour Segmentation in Whole Slide Images [12.961686610789416]
Self-supervised learning (SSL) has emerged as an alternative solution to reduce the annotation overheads in whole slide images.
These SSL approaches are not designed for handling multi-resolution WSIs, which limits their performance in learning discriminative image features.
We propose a Dual-branch SSL Framework for WSI tumour segmentation (DSF-WSI) that can effectively learn image features from multi-resolution WSIs.
arXiv Detail & Related papers (2023-03-20T10:57:28Z) - Learning Self-Supervised Low-Rank Network for Single-Stage Weakly and
Semi-Supervised Semantic Segmentation [119.009033745244]
This paper presents a Self-supervised Low-Rank Network ( SLRNet) for single-stage weakly supervised semantic segmentation (WSSS) and semi-supervised semantic segmentation (SSSS)
SLRNet uses cross-view self-supervision, that is, it simultaneously predicts several attentive LR representations from different views of an image to learn precise pseudo-labels.
Experiments on the Pascal VOC 2012, COCO, and L2ID datasets demonstrate that our SLRNet outperforms both state-of-the-art WSSS and SSSS methods with a variety of different settings.
arXiv Detail & Related papers (2022-03-19T09:19:55Z) - Semantic-Aware Generation for Self-Supervised Visual Representation
Learning [116.5814634936371]
We advocate for Semantic-aware Generation (SaGe) to facilitate richer semantics rather than details to be preserved in the generated image.
SaGe complements the target network with view-specific features and thus alleviates the semantic degradation brought by intensive data augmentations.
We execute SaGe on ImageNet-1K and evaluate the pre-trained models on five downstream tasks including nearest neighbor test, linear classification, and fine-scaled image recognition.
arXiv Detail & Related papers (2021-11-25T16:46:13Z) - Improving Semi-Supervised and Domain-Adaptive Semantic Segmentation with
Self-Supervised Depth Estimation [94.16816278191477]
We present a framework for semi-adaptive and domain-supervised semantic segmentation.
It is enhanced by self-supervised monocular depth estimation trained only on unlabeled image sequences.
We validate the proposed model on the Cityscapes dataset.
arXiv Detail & Related papers (2021-08-28T01:33:38Z) - Iterative, Deep, and Unsupervised Synthetic Aperture Sonar Image
Segmentation [29.435946984214937]
We present a new iterative unsupervised algorithm for learning deep features for SAS image segmentation.
Our results show that the performance of our proposed method is considerably better than current state-of-the-art methods in SAS image segmentation.
arXiv Detail & Related papers (2021-07-30T11:37:33Z) - AugNet: End-to-End Unsupervised Visual Representation Learning with
Image Augmentation [3.6790362352712873]
We propose AugNet, a new deep learning training paradigm to learn image features from a collection of unlabeled pictures.
Our experiments demonstrate that the method is able to represent the image in low dimensional space.
Unlike many deep-learning-based image retrieval algorithms, our approach does not require access to external annotated datasets.
arXiv Detail & Related papers (2021-06-11T09:02:30Z) - Large-scale Unsupervised Semantic Segmentation [163.3568726730319]
We propose a new problem of large-scale unsupervised semantic segmentation (LUSS) with a newly created benchmark dataset to track the research progress.
Based on the ImageNet dataset, we propose the ImageNet-S dataset with 1.2 million training images and 40k high-quality semantic segmentation annotations for evaluation.
arXiv Detail & Related papers (2021-06-06T15:02:11Z) - Structural Prior Driven Regularized Deep Learning for Sonar Image
Classification [28.306713374371814]
Deep learning has been shown to improve performance in the domain of synthetic aperture sonar (SAS) image classification.
Despite deep learning's recent success, there are still compelling open challenges in reducing the high false alarm rate.
We introduce a new deep learning architecture which incorporates priors with the goal of improving automatic target recognition.
arXiv Detail & Related papers (2020-10-26T04:00:46Z) - SCAN: Learning to Classify Images without Labels [73.69513783788622]
We advocate a two-step approach where feature learning and clustering are decoupled.
A self-supervised task from representation learning is employed to obtain semantically meaningful features.
We obtain promising results on ImageNet, and outperform several semi-supervised learning methods in the low-data regime.
arXiv Detail & Related papers (2020-05-25T18:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.