Advances in Self-Supervised Learning for Synthetic Aperture Sonar Data
Processing, Classification, and Pattern Recognition
- URL: http://arxiv.org/abs/2308.11633v1
- Date: Sat, 12 Aug 2023 20:59:39 GMT
- Title: Advances in Self-Supervised Learning for Synthetic Aperture Sonar Data
Processing, Classification, and Pattern Recognition
- Authors: Brandon Sheffield, Frank E. Bobe III, Bradley Marchand, Matthew S.
Emigh
- Abstract summary: This paper proposes MoCo-SAS that leverages self-supervised learning for SAS data processing, classification, and pattern recognition.
The experimental results demonstrate that MoCo-SAS significantly outperforms traditional supervised learning methods.
These findings highlight the potential of SSL in advancing the state-of-the-art in SAS data processing, offering promising avenues for enhanced underwater object detection and classification.
- Score: 0.36700088931938835
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Synthetic Aperture Sonar (SAS) imaging has become a crucial technology for
underwater exploration because of its unique ability to maintain resolution at
increasing ranges, a characteristic absent in conventional sonar techniques.
However, the effective application of deep learning to SAS data processing is
often limited due to the scarcity of labeled data. To address this challenge,
this paper proposes MoCo-SAS that leverages self-supervised learning (SSL) for
SAS data processing, classification, and pattern recognition. The experimental
results demonstrate that MoCo-SAS significantly outperforms traditional
supervised learning methods, as evidenced by significant improvements observed
in terms of the F1-score. These findings highlight the potential of SSL in
advancing the state-of-the-art in SAS data processing, offering promising
avenues for enhanced underwater object detection and classification.
Related papers
- Efficient Self-Supervised Learning for Earth Observation via Dynamic Dataset Curation [67.23953699167274]
Self-supervised learning (SSL) has enabled the development of vision foundation models for Earth Observation (EO)
In EO, this challenge is amplified by the redundancy and heavy-tailed distributions common in satellite imagery.
We propose a dynamic dataset pruning strategy designed to improve SSL pre-training by maximizing dataset diversity and balance.
arXiv Detail & Related papers (2025-04-09T15:13:26Z) - Incremental Self-training for Semi-supervised Learning [56.57057576885672]
IST is simple yet effective and fits existing self-training-based semi-supervised learning methods.
We verify the proposed IST on five datasets and two types of backbone, effectively improving the recognition accuracy and learning speed.
arXiv Detail & Related papers (2024-04-14T05:02:00Z) - Semi-Supervised Class-Agnostic Motion Prediction with Pseudo Label
Regeneration and BEVMix [59.55173022987071]
We study the potential of semi-supervised learning for class-agnostic motion prediction.
Our framework adopts a consistency-based self-training paradigm, enabling the model to learn from unlabeled data.
Our method exhibits comparable performance to weakly and some fully supervised methods.
arXiv Detail & Related papers (2023-12-13T09:32:50Z) - SASSL: Enhancing Self-Supervised Learning via Neural Style Transfer [20.769072160203038]
SASSL: Style Augmentations for Self Supervised Learning is a novel data augmentation technique based on Neural Style Transfer.
SASSL decouples semantic and stylistic attributes in images and applies exclusively to their style while preserving content.
SASSL boosts top-1 image classification accuracy on ImageNet by up to 2 percentage points compared to established self-supervised methods.
arXiv Detail & Related papers (2023-12-02T17:25:30Z) - Why Does Sharpness-Aware Minimization Generalize Better Than SGD? [102.40907275290891]
We show why Sharpness-Aware Minimization (SAM) generalizes better than Gradient Descent (SGD) for certain data model and two-layer convolutional ReLU networks.
Our result explains the benefits of SAM, particularly its ability to prevent noise learning in the early stages, thereby facilitating more effective learning of features.
arXiv Detail & Related papers (2023-10-11T07:51:10Z) - Sudden Drops in the Loss: Syntax Acquisition, Phase Transitions, and
Simplicity Bias in MLMs [50.5783641817253]
We present a case study of syntax acquisition in masked language models (MLMs)
We study Syntactic Attention Structure (SAS), a naturally emerging property of accessibles wherein specific Transformer heads tend to focus on specific syntactic relations.
We examine the causal role of SAS by manipulating SAS during training, and demonstrate that SAS is necessary for the development of grammatical capabilities.
arXiv Detail & Related papers (2023-09-13T20:57:11Z) - Self-Supervised Learning for Improved Synthetic Aperture Sonar Target
Recognition [0.0]
This study explores the application of self-supervised learning (SSL) for improved target recognition in synthetic aperture sonar (SAS) imagery.
The voluminous high-resolution SAS data presents a significant challenge for labeling; a crucial step for training deep neural networks (DNNs)
The study evaluates the performance of two prominent SSL algorithms, MoCov2 and BYOL, against the well-regarded supervised learning model, ResNet18, for binary image classification tasks.
arXiv Detail & Related papers (2023-07-27T14:17:24Z) - Iterative, Deep Synthetic Aperture Sonar Image Segmentation [21.319490900396474]
We propose an unsupervised learning framework called Iterative Deep Unsupervised (IDUS) for SAS image segmentation.
IDUS can be divided into four main steps: 1) A deep network estimates class assignments; 2) Low-level image features from the deep network are clustered into superpixels; 3) Superpixels are clustered into class assignments; 4) Resulting pseudo-labels are used for loss backpropagation of the deep network prediction.
A comparison of IDUS to current state-of-the-art methods on a realistic benchmark dataset for SAS image segmentation demonstrates the benefits of our proposal.
arXiv Detail & Related papers (2022-03-28T20:41:24Z) - Improved Speech Emotion Recognition using Transfer Learning and
Spectrogram Augmentation [56.264157127549446]
Speech emotion recognition (SER) is a challenging task that plays a crucial role in natural human-computer interaction.
One of the main challenges in SER is data scarcity.
We propose a transfer learning strategy combined with spectrogram augmentation.
arXiv Detail & Related papers (2021-08-05T10:39:39Z) - Structural Prior Driven Regularized Deep Learning for Sonar Image
Classification [28.306713374371814]
Deep learning has been shown to improve performance in the domain of synthetic aperture sonar (SAS) image classification.
Despite deep learning's recent success, there are still compelling open challenges in reducing the high false alarm rate.
We introduce a new deep learning architecture which incorporates priors with the goal of improving automatic target recognition.
arXiv Detail & Related papers (2020-10-26T04:00:46Z) - Uncovering the structure of clinical EEG signals with self-supervised
learning [64.4754948595556]
Supervised learning paradigms are often limited by the amount of labeled data that is available.
This phenomenon is particularly problematic in clinically-relevant data, such as electroencephalography (EEG)
By extracting information from unlabeled data, it might be possible to reach competitive performance with deep neural networks.
arXiv Detail & Related papers (2020-07-31T14:34:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.