Procrustean Training for Imbalanced Deep Learning
- URL: http://arxiv.org/abs/2104.01769v1
- Date: Mon, 5 Apr 2021 04:44:01 GMT
- Title: Procrustean Training for Imbalanced Deep Learning
- Authors: Han-Jia Ye, De-Chuan Zhan, Wei-Lun Chao
- Abstract summary: We show that a neural network tends to first under-fit the minor classes by classifying most of their data into the major classes.
We propose a novel learning strategy to equalize the training progress across classes.
- Score: 40.85940706868622
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural networks trained with class-imbalanced data are known to perform
poorly on minor classes of scarce training data. Several recent works attribute
this to over-fitting to minor classes. In this paper, we provide a novel
explanation of this issue. We found that a neural network tends to first
under-fit the minor classes by classifying most of their data into the major
classes in early training epochs. To correct these wrong predictions, the
neural network then must focus on pushing features of minor class data across
the decision boundaries between major and minor classes, leading to much larger
gradients for features of minor classes. We argue that such an under-fitting
phase over-emphasizes the competition between major and minor classes, hinders
the neural network from learning the discriminative knowledge that can be
generalized to test data, and eventually results in over-fitting. To address
this issue, we propose a novel learning strategy to equalize the training
progress across classes. We mix features of the major class data with those of
other data in a mini-batch, intentionally weakening their features to prevent a
neural network from fitting them first. We show that this strategy can largely
balance the training accuracy and feature gradients across classes, effectively
mitigating the under-fitting then over-fitting problem for minor class data. On
several benchmark datasets, our approach achieves the state-of-the-art
accuracy, especially for the challenging step-imbalanced cases.
Related papers
- Simplifying Neural Network Training Under Class Imbalance [77.39968702907817]
Real-world datasets are often highly class-imbalanced, which can adversely impact the performance of deep learning models.
The majority of research on training neural networks under class imbalance has focused on specialized loss functions, sampling techniques, or two-stage training procedures.
We demonstrate that simply tuning existing components of standard deep learning pipelines, such as the batch size, data augmentation, and label smoothing, can achieve state-of-the-art performance without any such specialized class imbalance methods.
arXiv Detail & Related papers (2023-12-05T05:52:44Z) - DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning [5.2319020651074215]
We propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP)
Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution.
We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors.
arXiv Detail & Related papers (2023-02-25T08:16:21Z) - Slimmable Networks for Contrastive Self-supervised Learning [69.9454691873866]
Self-supervised learning makes significant progress in pre-training large models, but struggles with small models.
We introduce another one-stage solution to obtain pre-trained small models without the need for extra teachers.
A slimmable network consists of a full network and several weight-sharing sub-networks, which can be pre-trained once to obtain various networks.
arXiv Detail & Related papers (2022-09-30T15:15:05Z) - Class Impression for Data-free Incremental Learning [20.23329169244367]
Deep learning-based classification approaches require collecting all samples from all classes in advance and are trained offline.
This paradigm may not be practical in real-world clinical applications, where new classes are incrementally introduced through the addition of new data.
We propose a novel data-free class incremental learning framework that first synthesizes data from the model trained on previous classes to generate a ours.
arXiv Detail & Related papers (2022-06-26T06:20:17Z) - Imbalanced Classification via Explicit Gradient Learning From Augmented
Data [0.0]
We propose a novel deep meta-learning technique to augment a given imbalanced dataset with new minority instances.
The advantage of the proposed method is demonstrated on synthetic and real-world datasets with various imbalance ratios.
arXiv Detail & Related papers (2022-02-21T22:16:50Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - CvS: Classification via Segmentation For Small Datasets [52.821178654631254]
This paper presents CvS, a cost-effective classifier for small datasets that derives the classification labels from predicting the segmentation maps.
We evaluate the effectiveness of our framework on diverse problems showing that CvS is able to achieve much higher classification results compared to previous methods when given only a handful of examples.
arXiv Detail & Related papers (2021-10-29T18:41:15Z) - Analyzing Overfitting under Class Imbalance in Neural Networks for Image
Segmentation [19.259574003403998]
In image segmentation neural networks may overfit to the foreground samples from small structures.
In this study, we provide new insights on the problem of overfitting under class imbalance by inspecting the network behavior.
arXiv Detail & Related papers (2021-02-20T14:57:58Z) - Identifying and Compensating for Feature Deviation in Imbalanced Deep
Learning [59.65752299209042]
We investigate learning a ConvNet under such a scenario.
We found that a ConvNet significantly over-fits the minor classes.
We propose to incorporate class-dependent temperatures (CDT) training ConvNet.
arXiv Detail & Related papers (2020-01-06T03:52:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.