A Feature Fusion-Net Using Deep Spatial Context Encoder and
Nonstationary Joint Statistical Model for High Resolution SAR Image
Classification
- URL: http://arxiv.org/abs/2105.04799v1
- Date: Tue, 11 May 2021 06:20:14 GMT
- Title: A Feature Fusion-Net Using Deep Spatial Context Encoder and
Nonstationary Joint Statistical Model for High Resolution SAR Image
Classification
- Authors: Wenkai Liang, Yan Wu, Ming Li, Peng Zhang, Yice Cao, Xin Hu
- Abstract summary: A novel end-to-end supervised classification method is proposed for HR SAR images.
To extract more effective spatial features, a new deep spatial context encoder network (DSCEN) is proposed.
To enhance the diversity of statistics, the nonstationary joint statistical model (NS-JSM) is adopted to form the global statistical features.
- Score: 10.152675581771113
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional neural networks (CNNs) have been applied to learn spatial
features for high-resolution (HR) synthetic aperture radar (SAR) image
classification. However, there has been little work on integrating the unique
statistical distributions of SAR images which can reveal physical properties of
terrain objects, into CNNs in a supervised feature learning framework. To
address this problem, a novel end-to-end supervised classification method is
proposed for HR SAR images by considering both spatial context and statistical
features. First, to extract more effective spatial features from SAR images, a
new deep spatial context encoder network (DSCEN) is proposed, which is a
lightweight structure and can be effectively trained with a small number of
samples. Meanwhile, to enhance the diversity of statistics, the nonstationary
joint statistical model (NS-JSM) is adopted to form the global statistical
features. Specifically, SAR images are transformed into the Gabor wavelet
domain and the produced multi-subbands magnitudes and phases are modeled by the
log-normal and uniform distribution. The covariance matrix is further utilized
to capture the inter-scale and intra-scale nonstationary correlation between
the statistical subbands and make the joint statistical features more compact
and distinguishable. Considering complementary advantages, a feature fusion
network (Fusion-Net) base on group compression and smooth normalization is
constructed to embed the statistical features into the spatial features and
optimize the fusion feature representation. As a result, our model can learn
the discriminative features and improve the final classification performance.
Experiments on four HR SAR images validate the superiority of the proposed
method over other related algorithms.
Related papers
- Rotated Multi-Scale Interaction Network for Referring Remote Sensing Image Segmentation [63.15257949821558]
Referring Remote Sensing Image (RRSIS) is a new challenge that combines computer vision and natural language processing.
Traditional Referring Image (RIS) approaches have been impeded by the complex spatial scales and orientations found in aerial imagery.
We introduce the Rotated Multi-Scale Interaction Network (RMSIN), an innovative approach designed for the unique demands of RRSIS.
arXiv Detail & Related papers (2023-12-19T08:14:14Z) - ESSAformer: Efficient Transformer for Hyperspectral Image
Super-resolution [76.7408734079706]
Single hyperspectral image super-resolution (single-HSI-SR) aims to restore a high-resolution hyperspectral image from a low-resolution observation.
We propose ESSAformer, an ESSA attention-embedded Transformer network for single-HSI-SR with an iterative refining structure.
arXiv Detail & Related papers (2023-07-26T07:45:14Z) - DCN-T: Dual Context Network with Transformer for Hyperspectral Image
Classification [109.09061514799413]
Hyperspectral image (HSI) classification is challenging due to spatial variability caused by complex imaging conditions.
We propose a tri-spectral image generation pipeline that transforms HSI into high-quality tri-spectral images.
Our proposed method outperforms state-of-the-art methods for HSI classification.
arXiv Detail & Related papers (2023-04-19T18:32:52Z) - Learning to Aggregate Multi-Scale Context for Instance Segmentation in
Remote Sensing Images [28.560068780733342]
A novel context aggregation network (CATNet) is proposed to improve the feature extraction process.
The proposed model exploits three lightweight plug-and-play modules, namely dense feature pyramid network (DenseFPN), spatial context pyramid ( SCP), and hierarchical region of interest extractor (HRoIE)
arXiv Detail & Related papers (2021-11-22T08:55:25Z) - Sparse Signal Models for Data Augmentation in Deep Learning ATR [0.8999056386710496]
We propose a data augmentation approach to incorporate domain knowledge and improve the generalization power of a data-intensive learning algorithm.
We exploit the sparsity of the scattering centers in the spatial domain and the smoothly-varying structure of the scattering coefficients in the azimuthal domain to solve the ill-posed problem of over-parametrized model fitting.
arXiv Detail & Related papers (2020-12-16T21:46:33Z) - Out-of-distribution Generalization via Partial Feature Decorrelation [72.96261704851683]
We present a novel Partial Feature Decorrelation Learning (PFDL) algorithm, which jointly optimize a feature decomposition network and the target image classification model.
The experiments on real-world datasets demonstrate that our method can improve the backbone model's accuracy on OOD image classification datasets.
arXiv Detail & Related papers (2020-07-30T05:48:48Z) - Multi-Objective CNN Based Algorithm for SAR Despeckling [1.933681537640272]
This paper proposes a convolutional neural network (CNN) with a multi-objective cost function taking care of SAR image properties.
Experiments on simulated and real SAR images show the accuracy of the proposed method compared to the State-of-Art despeckling algorithms.
arXiv Detail & Related papers (2020-06-16T10:15:42Z) - Image Super-Resolution with Cross-Scale Non-Local Attention and
Exhaustive Self-Exemplars Mining [66.82470461139376]
We propose the first Cross-Scale Non-Local (CS-NL) attention module with integration into a recurrent neural network.
By combining the new CS-NL prior with local and in-scale non-local priors in a powerful recurrent fusion cell, we can find more cross-scale feature correlations within a single low-resolution image.
arXiv Detail & Related papers (2020-06-02T07:08:58Z) - Spatial-Spectral Residual Network for Hyperspectral Image
Super-Resolution [82.1739023587565]
We propose a novel spectral-spatial residual network for hyperspectral image super-resolution (SSRNet)
Our method can effectively explore spatial-spectral information by using 3D convolution instead of 2D convolution, which enables the network to better extract potential information.
In each unit, we employ spatial and temporal separable 3D convolution to extract spatial and spectral information, which not only reduces unaffordable memory usage and high computational cost, but also makes the network easier to train.
arXiv Detail & Related papers (2020-01-14T03:34:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.