Towards Inter-class and Intra-class Imbalance in Class-imbalanced
Learning
- URL: http://arxiv.org/abs/2111.12791v1
- Date: Wed, 24 Nov 2021 20:50:54 GMT
- Title: Towards Inter-class and Intra-class Imbalance in Class-imbalanced
Learning
- Authors: Zhining Liu, Pengfei Wei, Zhepei Wei, Boyang Yu, Jing Jiang, Wei Cao,
Jiang Bian and Yi Chang
- Abstract summary: Imbalanced Learning (IL) is an important problem that widely exists in data mining applications.
We present Duple-Balanced Ensemble, namely DUBE, a versatile ensemble learning framework.
Unlike prevailing methods, DUBE directly performs inter-class and intra-class balancing without relying on heavy distance-based computation.
- Score: 24.01370257373491
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Imbalanced Learning (IL) is an important problem that widely exists in data
mining applications. Typical IL methods utilize intuitive class-wise resampling
or reweighting to directly balance the training set. However, some recent
research efforts in specific domains show that class-imbalanced learning can be
achieved without class-wise manipulation. This prompts us to think about the
relationship between the two different IL strategies and the nature of the
class imbalance. Fundamentally, they correspond to two essential imbalances
that exist in IL: the difference in quantity between examples from different
classes as well as between easy and hard examples within a single class, i.e.,
inter-class and intra-class imbalance. Existing works fail to explicitly take
both imbalances into account and thus suffer from suboptimal performance. In
light of this, we present Duple-Balanced Ensemble, namely DUBE , a versatile
ensemble learning framework. Unlike prevailing methods, DUBE directly performs
inter-class and intra-class balancing without relying on heavy distance-based
computation, which allows it to achieve competitive performance while being
computationally efficient. We also present a detailed discussion and analysis
about the pros and cons of different inter/intra-class balancing strategies
based on DUBE . Extensive experiments validate the effectiveness of the
proposed method. Code and examples are available at
https://github.com/ICDE2022Sub/duplebalance.
Related papers
- BaCon: Boosting Imbalanced Semi-supervised Learning via Balanced Feature-Level Contrastive Learning [0.9160375060389779]
In Class Imbalanced Semi-supervised Learning (CISSL), the bias introduced by unreliable pseudo-labels can be exacerbated by imbalanced data distributions.
Our method directly regularizes the distribution of instances' representations in a well-designed contrastive manner.
Our method demonstrates its effectiveness through comprehensive experiments on the CIFAR10-LT, CIFAR100-LT, STL10-LT, and SVHN-LT datasets.
arXiv Detail & Related papers (2024-03-04T06:43:16Z) - Simplifying Neural Network Training Under Class Imbalance [77.39968702907817]
Real-world datasets are often highly class-imbalanced, which can adversely impact the performance of deep learning models.
The majority of research on training neural networks under class imbalance has focused on specialized loss functions, sampling techniques, or two-stage training procedures.
We demonstrate that simply tuning existing components of standard deep learning pipelines, such as the batch size, data augmentation, and label smoothing, can achieve state-of-the-art performance without any such specialized class imbalance methods.
arXiv Detail & Related papers (2023-12-05T05:52:44Z) - Class Uncertainty: A Measure to Mitigate Class Imbalance [0.0]
We show that considering solely the cardinality of classes does not cover all issues causing class imbalance.
We propose "Class Uncertainty" as the average predictive uncertainty of the training examples.
We also curate SVCI-20 as a novel dataset in which the classes have equal number of training examples but they differ in terms of their hardness.
arXiv Detail & Related papers (2023-11-23T16:36:03Z) - Class-Imbalanced Graph Learning without Class Rebalancing [62.1368829847041]
Class imbalance is prevalent in real-world node classification tasks and poses great challenges for graph learning models.
In this work, we approach the root cause of class-imbalance bias from an topological paradigm.
We devise a lightweight topological augmentation framework BAT to mitigate the class-imbalance bias without class rebalancing.
arXiv Detail & Related papers (2023-08-27T19:01:29Z) - Rethinking Class Imbalance in Machine Learning [1.4467794332678536]
Imbalance learning is a subfield of machine learning that focuses on learning tasks in the presence of class imbalance.
This study presents a new taxonomy of class imbalance in machine learning with a broader scope.
We propose a new logit perturbation-based imbalance learning loss when proportion, variance, and distance imbalances exist simultaneously.
arXiv Detail & Related papers (2023-05-06T02:36:39Z) - An Embarrassingly Simple Baseline for Imbalanced Semi-Supervised
Learning [103.65758569417702]
Semi-supervised learning (SSL) has shown great promise in leveraging unlabeled data to improve model performance.
We consider a more realistic and challenging setting called imbalanced SSL, where imbalanced class distributions occur in both labeled and unlabeled data.
We study a simple yet overlooked baseline -- SimiS -- which tackles data imbalance by simply supplementing labeled data with pseudo-labels.
arXiv Detail & Related papers (2022-11-20T21:18:41Z) - Deep Reinforcement Learning for Multi-class Imbalanced Training [64.9100301614621]
We introduce an imbalanced classification framework, based on reinforcement learning, for training extremely imbalanced data sets.
We formulate a custom reward function and episode-training procedure, specifically with the added capability of handling multi-class imbalanced training.
Using real-world clinical case studies, we demonstrate that our proposed framework outperforms current state-of-the-art imbalanced learning methods.
arXiv Detail & Related papers (2022-05-24T13:39:59Z) - Neural Collapse Inspired Attraction-Repulsion-Balanced Loss for
Imbalanced Learning [97.81549071978789]
We propose Attraction-Repulsion-Balanced Loss (ARB-Loss) to balance the different components of the gradients.
We perform experiments on the large-scale classification and segmentation datasets and our ARB-Loss can achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-04-19T08:23:23Z) - Imbalanced Adversarial Training with Reweighting [33.51820466479575]
We show that adversarially trained models can suffer much worse performance on under-represented classes, when the training dataset is imbalanced.
Traditional reweighting strategies may lose efficacy to deal with the imbalance issue for adversarial training.
We propose Separable Reweighted Adversarial Training (SRAT) to facilitate adversarial training under imbalanced scenarios.
arXiv Detail & Related papers (2021-07-28T20:51:36Z) - Few-Shot Learning with Class Imbalance [13.60699610822265]
Few-shot learning aims to train models on a limited number of labeled samples given in a support set in order to generalize to unseen samples from a query set.
In the standard setup, the support set contains an equal amount of data points for each class.
We present a detailed study of few-shot class-imbalance along three axes: meta-dataset vs. task imbalance, effect of different imbalance distributions (linear, step, random), and effect of rebalancing techniques.
arXiv Detail & Related papers (2021-01-07T12:54:32Z) - Long-Tailed Recognition Using Class-Balanced Experts [128.73438243408393]
We propose an ensemble of class-balanced experts that combines the strength of diverse classifiers.
Our ensemble of class-balanced experts reaches results close to state-of-the-art and an extended ensemble establishes a new state-of-the-art on two benchmarks for long-tailed recognition.
arXiv Detail & Related papers (2020-04-07T20:57:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.