Pose Discrepancy Spatial Transformer Based Feature Disentangling for
Partial Aspect Angles SAR Target Recognition
- URL: http://arxiv.org/abs/2103.04329v1
- Date: Sun, 7 Mar 2021 11:47:34 GMT
- Title: Pose Discrepancy Spatial Transformer Based Feature Disentangling for
Partial Aspect Angles SAR Target Recognition
- Authors: Zaidao Wen, Jiaxiang Liu, Zhunga Liu, Quan Pan
- Abstract summary: This letter presents a novel framework termed DistSTN for the task of synthetic aperture radar (SAR) automatic target recognition (ATR)
In contrast to the conventional SAR ATR algorithms, DistSTN considers a more challenging practical scenario for non-cooperative targets.
We develop an amortized inference scheme that enables efficient feature extraction and recognition using an encoder-decoder mechanism.
- Score: 11.552273102567048
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This letter presents a novel framework termed DistSTN for the task of
synthetic aperture radar (SAR) automatic target recognition (ATR). In contrast
to the conventional SAR ATR algorithms, DistSTN considers a more challenging
practical scenario for non-cooperative targets whose aspect angles for training
are incomplete and limited in a partial range while those of testing samples
are unlimited. To address this issue, instead of learning the pose invariant
features, DistSTN newly involves an elaborated feature disentangling model to
separate the learned pose factors of a SAR target from the identity ones so
that they can independently control the representation process of the target
image. To disentangle the explainable pose factors, we develop a pose
discrepancy spatial transformer module in DistSTN to characterize the intrinsic
transformation between the factors of two different targets with an explicit
geometric model. Furthermore, DistSTN develops an amortized inference scheme
that enables efficient feature extraction and recognition using an
encoder-decoder mechanism. Experimental results with the moving and stationary
target acquisition and recognition (MSTAR) benchmark demonstrate the
effectiveness of our proposed approach. Compared with the other ATR algorithms,
DistSTN can achieve higher recognition accuracy.
Related papers
- PseudoNeg-MAE: Self-Supervised Point Cloud Learning using Conditional Pseudo-Negative Embeddings [55.55445978692678]
PseudoNeg-MAE is a self-supervised learning framework that enhances global feature representation of point cloud mask autoencoders.
We show that PseudoNeg-MAE achieves state-of-the-art performance on the ModelNet40 and ScanObjectNN datasets.
arXiv Detail & Related papers (2024-09-24T07:57:21Z) - Renormalized Connection for Scale-preferred Object Detection in Satellite Imagery [51.83786195178233]
We design a Knowledge Discovery Network (KDN) to implement the renormalization group theory in terms of efficient feature extraction.
Renormalized connection (RC) on the KDN enables synergistic focusing'' of multi-scale features.
RCs extend the multi-level feature's divide-and-conquer'' mechanism of the FPN-based detectors to a wide range of scale-preferred tasks.
arXiv Detail & Related papers (2024-09-09T13:56:22Z) - Continual-MAE: Adaptive Distribution Masked Autoencoders for Continual Test-Time Adaptation [49.827306773992376]
Continual Test-Time Adaptation (CTTA) is proposed to migrate a source pre-trained model to continually changing target distributions.
Our proposed method attains state-of-the-art performance in both classification and segmentation CTTA tasks.
arXiv Detail & Related papers (2023-12-19T15:34:52Z) - Rotated Multi-Scale Interaction Network for Referring Remote Sensing Image Segmentation [63.15257949821558]
Referring Remote Sensing Image (RRSIS) is a new challenge that combines computer vision and natural language processing.
Traditional Referring Image (RIS) approaches have been impeded by the complex spatial scales and orientations found in aerial imagery.
We introduce the Rotated Multi-Scale Interaction Network (RMSIN), an innovative approach designed for the unique demands of RRSIS.
arXiv Detail & Related papers (2023-12-19T08:14:14Z) - Masked Momentum Contrastive Learning for Zero-shot Semantic
Understanding [39.424931953675994]
Self-supervised pretraining (SSP) has emerged as a popular technique in machine learning, enabling the extraction of meaningful feature representations without labelled data.
This study endeavours to evaluate the effectiveness of pure self-supervised learning (SSL) techniques in computer vision tasks.
arXiv Detail & Related papers (2023-08-22T13:55:57Z) - Spatial Transform Decoupling for Oriented Object Detection [43.44237345360947]
Vision Transformers (ViTs) have achieved remarkable success in computer vision tasks.
We present a novel approach, termed Spatial Transform Decoupling (STD), providing a simple-yet-effective solution for oriented object detection with ViTs.
arXiv Detail & Related papers (2023-08-21T08:36:23Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - Domain-adaptive Person Re-identification without Cross-camera Paired
Samples [12.041823465553875]
Cross-camera pedestrian samples collected from long-distance scenes often have no positive samples.
It is extremely challenging to use cross-camera negative samples to achieve cross-region pedestrian identity matching.
A novel domain-adaptive person re-ID method that focuses on cross-camera consistent discriminative feature learning is proposed.
arXiv Detail & Related papers (2023-07-13T02:42:28Z) - Learning Invariant Representation via Contrastive Feature Alignment for
Clutter Robust SAR Target Recognition [10.993101256393679]
This letter proposes a solution called Contrastive Feature Alignment (CFA) to learn invariant representation for robust recognition.
CFA combines both classification and CWMSE losses to train the model jointly.
The proposed CFA combines both classification and CWMSE losses to train the model jointly, which allows for the progressive learning of invariant target representation.
arXiv Detail & Related papers (2023-04-04T12:35:33Z) - Toward Certified Robustness Against Real-World Distribution Shifts [65.66374339500025]
We train a generative model to learn perturbations from data and define specifications with respect to the output of the learned model.
A unique challenge arising from this setting is that existing verifiers cannot tightly approximate sigmoid activations.
We propose a general meta-algorithm for handling sigmoid activations which leverages classical notions of counter-example-guided abstraction refinement.
arXiv Detail & Related papers (2022-06-08T04:09:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.