Deep Similarity Learning Loss Functions in Data Transformation for Class
Imbalance
- URL: http://arxiv.org/abs/2312.10556v1
- Date: Sat, 16 Dec 2023 23:10:09 GMT
- Title: Deep Similarity Learning Loss Functions in Data Transformation for Class
Imbalance
- Authors: Damian Horna and Lango Mateusz and Jerzy Stefanowski
- Abstract summary: In this paper, we use deep neural networks to train new representations of multi-class data.
Our proposal modifies the distribution of features, i.e. the positions of examples in the learned embedded representation, and it does not modify the class sizes.
Experiments with popular multi-class imbalanced benchmark data sets and three classifiers showed the advantage of the proposed approach.
- Score: 2.693342141713236
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Improving the classification of multi-class imbalanced data is more difficult
than its two-class counterpart. In this paper, we use deep neural networks to
train new representations of tabular multi-class data. Unlike the typically
developed re-sampling pre-processing methods, our proposal modifies the
distribution of features, i.e. the positions of examples in the learned
embedded representation, and it does not modify the class sizes. To learn such
embedded representations we introduced various definitions of triplet loss
functions: the simplest one uses weights related to the degree of class
imbalance, while the next proposals are intended for more complex distributions
of examples and aim to generate a safe neighborhood of minority examples.
Similarly to the resampling approaches, after applying such preprocessing,
different classifiers can be trained on new representations. Experiments with
popular multi-class imbalanced benchmark data sets and three classifiers showed
the advantage of the proposed approach over popular pre-processing methods as
well as basic versions of neural networks with classical loss function
formulations.
Related papers
- Rethinking Classifier Re-Training in Long-Tailed Recognition: A Simple
Logits Retargeting Approach [102.0769560460338]
We develop a simple logits approach (LORT) without the requirement of prior knowledge of the number of samples per class.
Our method achieves state-of-the-art performance on various imbalanced datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018.
arXiv Detail & Related papers (2024-03-01T03:27:08Z) - Convolutional autoencoder-based multimodal one-class classification [80.52334952912808]
One-class classification refers to approaches of learning using data from a single class only.
We propose a deep learning one-class classification method suitable for multimodal data.
arXiv Detail & Related papers (2023-09-25T12:31:18Z) - Generalization Bounds for Few-Shot Transfer Learning with Pretrained
Classifiers [26.844410679685424]
We study the ability of foundation models to learn representations for classification that are transferable to new, unseen classes.
We show that the few-shot error of the learned feature map on new classes is small in case of class-feature-variability collapse.
arXiv Detail & Related papers (2022-12-23T18:46:05Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Learning Debiased and Disentangled Representations for Semantic
Segmentation [52.35766945827972]
We propose a model-agnostic and training scheme for semantic segmentation.
By randomly eliminating certain class information in each training iteration, we effectively reduce feature dependencies among classes.
Models trained with our approach demonstrate strong results on multiple semantic segmentation benchmarks.
arXiv Detail & Related papers (2021-10-31T16:15:09Z) - SetConv: A New Approach for Learning from Imbalanced Data [29.366843553056594]
We propose a set convolution operation and an episodic training strategy to extract a single representative for each class.
We prove that our proposed algorithm is permutation-invariant despite the order of inputs.
arXiv Detail & Related papers (2021-04-03T22:33:30Z) - A Novel Adaptive Minority Oversampling Technique for Improved
Classification in Data Imbalanced Scenarios [23.257891827728827]
Imbalance in the proportion of training samples belonging to different classes often poses performance degradation of conventional classifiers.
We propose a novel three step technique to address imbalanced data.
arXiv Detail & Related papers (2021-03-24T09:58:02Z) - Variational Feature Disentangling for Fine-Grained Few-Shot
Classification [30.350307891161865]
Fine-grained few-shot recognition often suffers from the problem of training data scarcity for novel categories.
In this paper, we focus one enlarging the intra-class variance of the unseen class to improve few-shot classification performance.
arXiv Detail & Related papers (2020-10-07T08:13:42Z) - M2m: Imbalanced Classification via Major-to-minor Translation [79.09018382489506]
In most real-world scenarios, labeled training datasets are highly class-imbalanced, where deep neural networks suffer from generalizing to a balanced testing criterion.
In this paper, we explore a novel yet simple way to alleviate this issue by augmenting less-frequent classes via translating samples from more-frequent classes.
Our experimental results on a variety of class-imbalanced datasets show that the proposed method improves the generalization on minority classes significantly compared to other existing re-sampling or re-weighting methods.
arXiv Detail & Related papers (2020-04-01T13:21:17Z) - Selecting Relevant Features from a Multi-domain Representation for
Few-shot Classification [91.67977602992657]
We propose a new strategy based on feature selection, which is both simpler and more effective than previous feature adaptation approaches.
We show that a simple non-parametric classifier built on top of such features produces high accuracy and generalizes to domains never seen during training.
arXiv Detail & Related papers (2020-03-20T15:44:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.