Mining Minority-class Examples With Uncertainty Estimates
- URL: http://arxiv.org/abs/2112.07835v1
- Date: Wed, 15 Dec 2021 02:05:02 GMT
- Title: Mining Minority-class Examples With Uncertainty Estimates
- Authors: Gursimran Singh, Lingyang Chu, Lanjun Wang, Jian Pei, Qi Tian, Yong
Zhang
- Abstract summary: In the real world, the frequency of occurrence of objects is naturally skewed forming long-tail class distributions.
We propose an effective, yet simple, approach to overcome these challenges.
Our framework enhances the subdued tail-class activations and, thereafter, uses a one-class data-centric approach to effectively identify tail-class examples.
- Score: 102.814407678425
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the real world, the frequency of occurrence of objects is naturally skewed
forming long-tail class distributions, which results in poor performance on the
statistically rare classes. A promising solution is to mine tail-class examples
to balance the training dataset. However, mining tail-class examples is a very
challenging task. For instance, most of the otherwise successful
uncertainty-based mining approaches struggle due to distortion of class
probabilities resulting from skewness in data. In this work, we propose an
effective, yet simple, approach to overcome these challenges. Our framework
enhances the subdued tail-class activations and, thereafter, uses a one-class
data-centric approach to effectively identify tail-class examples. We carry out
an exhaustive evaluation of our framework on three datasets spanning over two
computer vision tasks. Substantial improvements in the minority-class mining
and fine-tuned model's performance strongly corroborate the value of our
proposed solution.
Related papers
- Better (pseudo-)labels for semi-supervised instance segmentation [21.703173564795353]
We introduce a dual-strategy to enhance the teacher model's training process, substantially improving the performance on few-shot learning.
We observe marked improvements over a state-of-the-art supervised baseline performance on the LVIS dataset, with an increase of 2.8% in average precision (AP) and 10.3% gain in AP for rare classes.
arXiv Detail & Related papers (2024-03-18T11:23:02Z) - Towards Accelerated Model Training via Bayesian Data Selection [45.62338106716745]
We propose a more reasonable data selection principle by examining the data's impact on the model's generalization loss.
Recent work has proposed a more reasonable data selection principle by examining the data's impact on the model's generalization loss.
This work solves these problems by leveraging a lightweight Bayesian treatment and incorporating off-the-shelf zero-shot predictors built on large-scale pre-trained models.
arXiv Detail & Related papers (2023-08-21T07:58:15Z) - Neural Collapse Terminus: A Unified Solution for Class Incremental
Learning and Its Variants [166.916517335816]
In this paper, we offer a unified solution to the misalignment dilemma in the three tasks.
We propose neural collapse terminus that is a fixed structure with the maximal equiangular inter-class separation for the whole label space.
Our method holds the neural collapse optimality in an incremental fashion regardless of data imbalance or data scarcity.
arXiv Detail & Related papers (2023-08-03T13:09:59Z) - Two-Stage Fine-Tuning: A Novel Strategy for Learning Class-Imbalanced
Data [11.66734752179563]
Classification on long-tailed distributed data is a challenging problem.
Learning on tail classes is especially challenging for the fine-tuning when transferring a pretrained model to a downstream task.
We propose a two-stage fine-tuning: we first fine-tune the final layer of the pretrained model with class-balanced reweighting loss, and then we perform the standard fine-tuning.
arXiv Detail & Related papers (2022-07-22T03:39:51Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - The Devil is in Classification: A Simple Framework for Long-tail Object
Detection and Instance Segmentation [93.17367076148348]
We investigate performance drop of the state-of-the-art two-stage instance segmentation model Mask R-CNN on the recent long-tail LVIS dataset.
We unveil that a major cause is the inaccurate classification of object proposals.
We propose a simple calibration framework to more effectively alleviate classification head bias with a bi-level class balanced sampling approach.
arXiv Detail & Related papers (2020-07-23T12:49:07Z) - Long-Tailed Recognition Using Class-Balanced Experts [128.73438243408393]
We propose an ensemble of class-balanced experts that combines the strength of diverse classifiers.
Our ensemble of class-balanced experts reaches results close to state-of-the-art and an extended ensemble establishes a new state-of-the-art on two benchmarks for long-tailed recognition.
arXiv Detail & Related papers (2020-04-07T20:57:44Z) - Robust and On-the-fly Dataset Denoising for Image Classification [72.10311040730815]
On-the-fly Data Denoising (ODD) is robust to mislabeled examples, while introducing almost zero computational overhead compared to standard training.
ODD is able to achieve state-of-the-art results on a wide range of datasets including real-world ones such as WebVision and Clothing1M.
arXiv Detail & Related papers (2020-03-24T03:59:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.