Impact of PolSAR pre-processing and balancing methods on complex-valued
neural networks segmentation tasks
- URL: http://arxiv.org/abs/2210.17419v1
- Date: Fri, 28 Oct 2022 12:49:43 GMT
- Title: Impact of PolSAR pre-processing and balancing methods on complex-valued
neural networks segmentation tasks
- Authors: Jos\'e Agustin Barrachina, Chengfang Ren, Christ\`ele Morisseau,
Gilles Vieillard, Jean-Philippe Ovarlez
- Abstract summary: We investigate the semantic segmentation of Polarimetric Synthetic Aperture Radar (PolSAR) using Complex-Valued Neural Network (CVNN)
We exhaustively compare both methods for six model architectures, three complex-valued, and their respective real-equivalent models.
We propose two methods for reducing this gap and performing the results for all input representations, models, and dataset pre-processing.
- Score: 9.6556424340252
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we investigated the semantic segmentation of Polarimetric
Synthetic Aperture Radar (PolSAR) using Complex-Valued Neural Network (CVNN).
Although the coherency matrix is more widely used as the input of CVNN, the
Pauli vector has recently been shown to be a valid alternative. We exhaustively
compare both methods for six model architectures, three complex-valued, and
their respective real-equivalent models. We are comparing, therefore, not only
the input representation impact but also the complex- against the real-valued
models. We then argue that the dataset splitting produces a high correlation
between training and validation sets, saturating the task and thus achieving
very high performance. We, therefore, use a different data pre-processing
technique designed to reduce this effect and reproduce the results with the
same configurations as before (input representation and model architectures).
After seeing that the performance per class is highly different according to
class occurrences, we propose two methods for reducing this gap and performing
the results for all input representations, models, and dataset pre-processing.
Related papers
- Representation Similarity: A Better Guidance of DNN Layer Sharing for Edge Computing without Training [3.792729116385123]
We propose a new model merging scheme by sharing representations at the edge, guided by representation similarity S.
We show that S is extremely highly correlated with merged model's accuracy with Pearson Correlation Coefficient |r| > 0.94 than other metrics.
arXiv Detail & Related papers (2024-10-15T03:35:54Z) - On Characterizing the Evolution of Embedding Space of Neural Networks
using Algebraic Topology [9.537910170141467]
We study how the topology of feature embedding space changes as it passes through the layers of a well-trained deep neural network (DNN) through Betti numbers.
We demonstrate that as depth increases, a topologically complicated dataset is transformed into a simple one, resulting in Betti numbers attaining their lowest possible value.
arXiv Detail & Related papers (2023-11-08T10:45:12Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Improved Distribution Matching for Dataset Condensation [91.55972945798531]
We propose a novel dataset condensation method based on distribution matching.
Our simple yet effective method outperforms most previous optimization-oriented methods with much fewer computational resources.
arXiv Detail & Related papers (2023-07-19T04:07:33Z) - Batch-Ensemble Stochastic Neural Networks for Out-of-Distribution
Detection [55.028065567756066]
Out-of-distribution (OOD) detection has recently received much attention from the machine learning community due to its importance in deploying machine learning models in real-world applications.
In this paper we propose an uncertainty quantification approach by modelling the distribution of features.
We incorporate an efficient ensemble mechanism, namely batch-ensemble, to construct the batch-ensemble neural networks (BE-SNNs) and overcome the feature collapse problem.
We show that BE-SNNs yield superior performance on several OOD benchmarks, such as the Two-Moons dataset, the FashionMNIST vs MNIST dataset, FashionM
arXiv Detail & Related papers (2022-06-26T16:00:22Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - SHRIMP: Sparser Random Feature Models via Iterative Magnitude Pruning [3.775565013663731]
We propose a new method -- Sparser Random Feature Models via IMP (ShRIMP) -- to efficiently fit high-dimensional data with inherent low-dimensional structure.
Our method can be viewed as a combined process to construct and find sparse lottery tickets for two-layer dense networks.
arXiv Detail & Related papers (2021-12-07T21:32:28Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - Complex-Valued vs. Real-Valued Neural Networks for Classification
Perspectives: An Example on Non-Circular Data [10.06162739966462]
We show the potential interest of Complex-Valued Neural Network (CVNN) on classification tasks for complex-valued datasets.
CVNN accuracy presents a statistically higher mean and median variance and lower variance than Real-Valued Neural Network (RVNN)
arXiv Detail & Related papers (2020-09-17T14:39:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.