Few-Shot Learning with Class Imbalance
- URL: http://arxiv.org/abs/2101.02523v1
- Date: Thu, 7 Jan 2021 12:54:32 GMT
- Title: Few-Shot Learning with Class Imbalance
- Authors: Mateusz Ochal, Massimiliano Patacchiola, Amos Storkey, Jose Vazquez,
Sen Wang
- Abstract summary: Few-shot learning aims to train models on a limited number of labeled samples given in a support set in order to generalize to unseen samples from a query set.
In the standard setup, the support set contains an equal amount of data points for each class.
We present a detailed study of few-shot class-imbalance along three axes: meta-dataset vs. task imbalance, effect of different imbalance distributions (linear, step, random), and effect of rebalancing techniques.
- Score: 13.60699610822265
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot learning aims to train models on a limited number of labeled samples
given in a support set in order to generalize to unseen samples from a query
set. In the standard setup, the support set contains an equal amount of data
points for each class. However, this assumption overlooks many practical
considerations arising from the dynamic nature of the real world, such as
class-imbalance. In this paper, we present a detailed study of few-shot
class-imbalance along three axes: meta-dataset vs. task imbalance, effect of
different imbalance distributions (linear, step, random), and effect of
rebalancing techniques. We extensively compare over 10 state-of-the-art
few-shot learning and meta-learning methods using unbalanced tasks and
meta-datasets. Our analysis using Mini-ImageNet reveals that 1) compared to the
balanced task, the performances on class-imbalance tasks counterparts always
drop, by up to $18.0\%$ for optimization-based methods, and up to $8.4$ for
metric-based methods, 2) contrary to popular belief, meta-learning algorithms,
such as MAML, do not automatically learn to balance by being exposed to
imbalanced tasks during (meta-)training time, 3) strategies used to mitigate
imbalance in supervised learning, such as oversampling, can offer a stronger
solution to the class imbalance problem, 4) the effect of imbalance at the
meta-dataset level is less significant than the effect at the task level with
similar imbalance magnitude. The code to reproduce the experiments is released
under an open-source license.
Related papers
- Simplifying Neural Network Training Under Class Imbalance [77.39968702907817]
Real-world datasets are often highly class-imbalanced, which can adversely impact the performance of deep learning models.
The majority of research on training neural networks under class imbalance has focused on specialized loss functions, sampling techniques, or two-stage training procedures.
We demonstrate that simply tuning existing components of standard deep learning pipelines, such as the batch size, data augmentation, and label smoothing, can achieve state-of-the-art performance without any such specialized class imbalance methods.
arXiv Detail & Related papers (2023-12-05T05:52:44Z) - Class-Imbalanced Graph Learning without Class Rebalancing [62.1368829847041]
Class imbalance is prevalent in real-world node classification tasks and poses great challenges for graph learning models.
In this work, we approach the root cause of class-imbalance bias from an topological paradigm.
We devise a lightweight topological augmentation framework BAT to mitigate the class-imbalance bias without class rebalancing.
arXiv Detail & Related papers (2023-08-27T19:01:29Z) - Rethinking Class Imbalance in Machine Learning [1.4467794332678536]
Imbalance learning is a subfield of machine learning that focuses on learning tasks in the presence of class imbalance.
This study presents a new taxonomy of class imbalance in machine learning with a broader scope.
We propose a new logit perturbation-based imbalance learning loss when proportion, variance, and distance imbalances exist simultaneously.
arXiv Detail & Related papers (2023-05-06T02:36:39Z) - Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification [74.62203971625173]
Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
arXiv Detail & Related papers (2022-08-05T01:23:54Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Minority Class Oriented Active Learning for Imbalanced Datasets [6.009262446889319]
We introduce a new active learning method which is designed for imbalanced datasets.
It favors samples likely to be in minority classes so as to reduce the imbalance of the labeled subset.
We also compare two training schemes for active learning.
arXiv Detail & Related papers (2022-02-01T13:13:41Z) - Towards Inter-class and Intra-class Imbalance in Class-imbalanced
Learning [24.01370257373491]
Imbalanced Learning (IL) is an important problem that widely exists in data mining applications.
We present Duple-Balanced Ensemble, namely DUBE, a versatile ensemble learning framework.
Unlike prevailing methods, DUBE directly performs inter-class and intra-class balancing without relying on heavy distance-based computation.
arXiv Detail & Related papers (2021-11-24T20:50:54Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - MetaBalance: High-Performance Neural Networks for Class-Imbalanced Data [42.81296448544681]
Class-imbalanced data, in which some classes contain far more samples than others, is ubiquitous in real-world applications.
Standard techniques for handling class-imbalance usually work by training on a re-weighted loss or on re-balanced data.
We evaluate our method, MetaBalance, on image classification, credit-card fraud detection, loan default prediction, and facial recognition tasks with severely imbalanced data.
arXiv Detail & Related papers (2021-06-17T16:42:50Z) - How Sensitive are Meta-Learners to Dataset Imbalance? [13.60699610822265]
We show that ML methods are more robust against meta-dataset imbalance than imbalance at the task-level.
Overall, these results highlight an implicit strength of ML algorithms, capable of learning generalizable features under dataset imbalance and domain-shift.
arXiv Detail & Related papers (2021-04-12T10:47:42Z) - Long-Tailed Recognition Using Class-Balanced Experts [128.73438243408393]
We propose an ensemble of class-balanced experts that combines the strength of diverse classifiers.
Our ensemble of class-balanced experts reaches results close to state-of-the-art and an extended ensemble establishes a new state-of-the-art on two benchmarks for long-tailed recognition.
arXiv Detail & Related papers (2020-04-07T20:57:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.