Exploring Deep Learning Methods for Classification of SAR Images:
Towards NextGen Convolutions via Transformers
- URL: http://arxiv.org/abs/2303.15852v1
- Date: Tue, 28 Mar 2023 09:43:58 GMT
- Title: Exploring Deep Learning Methods for Classification of SAR Images:
Towards NextGen Convolutions via Transformers
- Authors: Aakash Singh and Vivek Kumar Singh
- Abstract summary: This study is an attempt to explore the suitability of current state-of-the-art models introduced in the domain of computer vision for SAR target classification (MSTAR)
Experimental results show that deep learning models can be suitably applied in the domain of SAR image classification with the desired performance levels.
- Score: 1.8532775355974984
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Images generated by high-resolution SAR have vast areas of application as
they can work better in adverse light and weather conditions. One such area of
application is in the military systems. This study is an attempt to explore the
suitability of current state-of-the-art models introduced in the domain of
computer vision for SAR target classification (MSTAR). Since the application of
any solution produced for military systems would be strategic and real-time,
accuracy is often not the only criterion to measure its performance. Other
important parameters like prediction time and input resiliency are equally
important. The paper deals with these issues in the context of SAR images.
Experimental results show that deep learning models can be suitably applied in
the domain of SAR image classification with the desired performance levels.
Related papers
- SAR-W-MixMAE: SAR Foundation Model Training Using Backscatter Power Weighting [3.618534280726541]
Foundation model approaches such as masked auto-encoders (MAE) or its variations are now being successfully applied to satellite imagery.
Due to difficulty in semantic labeling to create datasets and higher noise content with respect to optical images, Synthetic Aperture Radar (SAR) data has not been explored a lot in the field for foundation models.
In this work, we explored masked auto-encoder, specifically MixMAE on Sentinel-1 SAR images and its impact on SAR image classification tasks.
arXiv Detail & Related papers (2025-03-03T05:09:44Z) - Enhancing SAR Object Detection with Self-Supervised Pre-training on Masked Auto-Encoders [5.234109158596138]
Self-supervised learning (SSL) is proposed to learn feature representations of SAR images during the pre-training process.
The proposed method captures proper latent representations of SAR images and improves the model generalization in downstream tasks.
arXiv Detail & Related papers (2025-01-20T03:28:34Z) - LiDAR-GS:Real-time LiDAR Re-Simulation using Gaussian Splatting [50.808933338389686]
We present LiDAR-GS, a real-time, high-fidelity re-simulation of LiDAR scans in public urban road scenes.
The method achieves state-of-the-art results in both rendering frame rate and quality on publically available large scene datasets.
arXiv Detail & Related papers (2024-10-07T15:07:56Z) - SAFE: a SAR Feature Extractor based on self-supervised learning and masked Siamese ViTs [5.961207817077044]
We propose a novel self-supervised learning framework based on masked Siamese Vision Transformers to create a General SAR Feature Extractor coined SAFE.
Our method leverages contrastive learning principles to train a model on unlabeled SAR data, extracting robust and generalizable features.
We introduce tailored data augmentation techniques specific to SAR imagery, such as sub-aperture decomposition and despeckling.
Our network competes with or surpasses other state-of-the-art methods in few-shot classification and segmentation tasks, even without being trained on the sensors used for the evaluation.
arXiv Detail & Related papers (2024-06-30T23:11:20Z) - Efficient Visual State Space Model for Image Deblurring [83.57239834238035]
Convolutional neural networks (CNNs) and Vision Transformers (ViTs) have achieved excellent performance in image restoration.
We propose a simple yet effective visual state space model (EVSSM) for image deblurring.
arXiv Detail & Related papers (2024-05-23T09:13:36Z) - Towards SAR Automatic Target Recognition MultiCategory SAR Image Classification Based on Light Weight Vision Transformer [11.983317593939688]
This paper tries to apply a lightweight vision transformer based model to classify SAR images.
The entire structure was verified by an open-accessed SAR data set.
arXiv Detail & Related papers (2024-05-18T11:24:52Z) - Efficient Prompt Tuning of Large Vision-Language Model for Fine-Grained
Ship Classification [62.425462136772666]
Fine-grained ship classification in remote sensing (RS-FGSC) poses a significant challenge due to the high similarity between classes and the limited availability of labeled data.
Recent advancements in large pre-trained Vision-Language Models (VLMs) have demonstrated impressive capabilities in few-shot or zero-shot learning.
This study delves into harnessing the potential of VLMs to enhance classification accuracy for unseen ship categories.
arXiv Detail & Related papers (2024-03-13T05:48:58Z) - Improved Difference Images for Change Detection Classifiers in SAR
Imagery Using Deep Learning [0.0]
This paper proposes a new method of improving SAR image processing to produce higher quality difference images for the classification algorithms.
The method is built on a neural network-based mapping transformation function that produces artificial SAR images from a location in the requested acquisition conditions.
arXiv Detail & Related papers (2023-03-31T06:57:34Z) - Remote Sensing Image Classification using Transfer Learning and
Attention Based Deep Neural Network [59.86658316440461]
We propose a deep learning based framework for RSISC, which makes use of the transfer learning technique and multihead attention scheme.
The proposed deep learning framework is evaluated on the benchmark NWPU-RESISC45 dataset and achieves the best classification accuracy of 94.7%.
arXiv Detail & Related papers (2022-06-20T10:05:38Z) - Textural-Structural Joint Learning for No-Reference Super-Resolution
Image Quality Assessment [59.91741119995321]
We develop a dual stream network to jointly explore the textural and structural information for quality prediction, dubbed TSNet.
By mimicking the human vision system (HVS) that pays more attention to the significant areas of the image, we develop the spatial attention mechanism to make the visual-sensitive areas more distinguishable.
Experimental results show the proposed TSNet predicts the visual quality more accurate than the state-of-the-art IQA methods, and demonstrates better consistency with the human's perspective.
arXiv Detail & Related papers (2022-05-27T09:20:06Z) - A Contrastive Learning Approach to Auroral Identification and
Classification [0.8399688944263843]
We present a novel application of unsupervised learning to the task of auroral image classification.
We modify and adapt the Simple framework for Contrastive Learning of Representations (SimCLR) algorithm to learn representations of auroral images.
Our approach exceeds an established threshold for operational purposes, demonstrating readiness for deployment and utilization.
arXiv Detail & Related papers (2021-09-28T17:51:25Z) - Cycle and Semantic Consistent Adversarial Domain Adaptation for Reducing
Simulation-to-Real Domain Shift in LiDAR Bird's Eye View [110.83289076967895]
We present a BEV domain adaptation method based on CycleGAN that uses prior semantic classification in order to preserve the information of small objects of interest during the domain adaptation process.
The quality of the generated BEVs has been evaluated using a state-of-the-art 3D object detection framework at KITTI 3D Object Detection Benchmark.
arXiv Detail & Related papers (2021-04-22T12:47:37Z) - PeaceGAN: A GAN-based Multi-Task Learning Method for SAR Target Image
Generation with a Pose Estimator and an Auxiliary Classifier [50.17500790309477]
We propose a novel GAN-based multi-task learning (MTL) method for SAR target image generation, called PeaceGAN.
PeaceGAN uses both pose angle and target class information, which makes it possible to produce SAR target images of desired target classes at intended pose angles.
arXiv Detail & Related papers (2021-03-29T10:03:09Z) - Visualization of Deep Transfer Learning In SAR Imagery [0.0]
We consider transfer learning to leverage deep features from a network trained on an EO ships dataset.
By exploring the network activations in the form of class-activation maps, we gain insight on how a deep network interprets a new modality.
arXiv Detail & Related papers (2021-03-20T00:16:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.