CUDA: Curriculum of Data Augmentation for Long-Tailed Recognition
- URL: http://arxiv.org/abs/2302.05499v1
- Date: Fri, 10 Feb 2023 20:30:22 GMT
- Title: CUDA: Curriculum of Data Augmentation for Long-Tailed Recognition
- Authors: Sumyeong Ahn, Jongwoo Ko, Se-Young Yun
- Abstract summary: Class imbalance problems frequently occur in real-world tasks.
To mitigate this problem, many approaches have aimed to balance among given classes by re-weighting or re-sampling training samples.
These re-balancing methods increase the impact of minority classes and reduce the influence of majority classes on the output of models.
Several methods have been developed that increase the representations of minority samples by the features of the majority samples.
- Score: 10.441880303257468
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Class imbalance problems frequently occur in real-world tasks, and
conventional deep learning algorithms are well known for performance
degradation on imbalanced training datasets. To mitigate this problem, many
approaches have aimed to balance among given classes by re-weighting or
re-sampling training samples. These re-balancing methods increase the impact of
minority classes and reduce the influence of majority classes on the output of
models. However, the extracted representations may be of poor quality owing to
the limited number of minority samples. To handle this restriction, several
methods have been developed that increase the representations of minority
samples by leveraging the features of the majority samples. Despite extensive
recent studies, no deep analysis has been conducted on determination of classes
to be augmented and strength of augmentation has been conducted. In this study,
we first investigate the correlation between the degree of augmentation and
class-wise performance, and find that the proper degree of augmentation must be
allocated for each class to mitigate class imbalance problems. Motivated by
this finding, we propose a simple and efficient novel curriculum, which is
designed to find the appropriate per-class strength of data augmentation,
called CUDA: CUrriculum of Data Augmentation for long-tailed recognition. CUDA
can simply be integrated into existing long-tailed recognition methods. We
present the results of experiments showing that CUDA effectively achieves
better generalization performance compared to the state-of-the-art method on
various imbalanced datasets such as CIFAR-100-LT, ImageNet-LT, and iNaturalist
2018.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - An Efficient Replay for Class-Incremental Learning with Pre-trained Models [0.0]
In class-incremental learning, the steady state among the weight guided by each class center is disrupted, which is significantly correlated with forgetting.
We propose a new method to overcoming forgetting.
arXiv Detail & Related papers (2024-08-15T11:26:28Z) - Rethinking Classifier Re-Training in Long-Tailed Recognition: A Simple
Logits Retargeting Approach [102.0769560460338]
We develop a simple logits approach (LORT) without the requirement of prior knowledge of the number of samples per class.
Our method achieves state-of-the-art performance on various imbalanced datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018.
arXiv Detail & Related papers (2024-03-01T03:27:08Z) - Combating Representation Learning Disparity with Geometric Harmonization [50.29859682439571]
We propose a novel Geometric Harmonization (GH) method to encourage category-level uniformity in representation learning.
Our proposal does not alter the setting of SSL and can be easily integrated into existing methods in a low-cost manner.
arXiv Detail & Related papers (2023-10-26T17:41:11Z) - A Unified Generalization Analysis of Re-Weighting and Logit-Adjustment
for Imbalanced Learning [129.63326990812234]
We propose a technique named data-dependent contraction to capture how modified losses handle different classes.
On top of this technique, a fine-grained generalization bound is established for imbalanced learning, which helps reveal the mystery of re-weighting and logit-adjustment.
arXiv Detail & Related papers (2023-10-07T09:15:08Z) - Generative Oversampling for Imbalanced Data via Majority-Guided VAE [15.93867386081279]
We propose a novel over-sampling model, called Majority-Guided VAE(MGVAE), which generates new minority samples under the guidance of a majority-based prior.
In this way, the newly generated minority samples can inherit the diversity and richness of the majority ones, thus mitigating overfitting in downstream tasks.
arXiv Detail & Related papers (2023-02-14T06:35:23Z) - Imbalanced Classification via Explicit Gradient Learning From Augmented
Data [0.0]
We propose a novel deep meta-learning technique to augment a given imbalanced dataset with new minority instances.
The advantage of the proposed method is demonstrated on synthetic and real-world datasets with various imbalance ratios.
arXiv Detail & Related papers (2022-02-21T22:16:50Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z) - Minority Class Oversampling for Tabular Data with Deep Generative Models [4.976007156860967]
We study the ability of deep generative models to provide realistic samples that improve performance on imbalanced classification tasks via oversampling.
Our experiments show that the way the method of sampling does not affect quality, but runtime varies widely.
We also observe that the improvements in terms of performance metric, while shown to be significant, often are minor in absolute terms.
arXiv Detail & Related papers (2020-05-07T21:35:57Z) - M2m: Imbalanced Classification via Major-to-minor Translation [79.09018382489506]
In most real-world scenarios, labeled training datasets are highly class-imbalanced, where deep neural networks suffer from generalizing to a balanced testing criterion.
In this paper, we explore a novel yet simple way to alleviate this issue by augmenting less-frequent classes via translating samples from more-frequent classes.
Our experimental results on a variety of class-imbalanced datasets show that the proposed method improves the generalization on minority classes significantly compared to other existing re-sampling or re-weighting methods.
arXiv Detail & Related papers (2020-04-01T13:21:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.