Reliable Label Correction is a Good Booster When Learning with Extremely
Noisy Labels
- URL: http://arxiv.org/abs/2205.00186v1
- Date: Sat, 30 Apr 2022 07:19:03 GMT
- Title: Reliable Label Correction is a Good Booster When Learning with Extremely
Noisy Labels
- Authors: Kai Wang, Xiangyu Peng, Shuo Yang, Jianfei Yang, Zheng Zhu, Xinchao
Wang and Yang You
- Abstract summary: We introduce a novel framework, termed as LC-Booster, to explicitly tackle learning under extreme noise.
LC-Booster incorporates label correction into the sample selection, so that more purified samples, through the reliable label correction, can be utilized for training.
Experiments show that LC-Booster advances state-of-the-art results on several noisy-label benchmarks.
- Score: 65.79898033530408
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning with noisy labels has aroused much research interest since data
annotations, especially for large-scale datasets, may be inevitably imperfect.
Recent approaches resort to a semi-supervised learning problem by dividing
training samples into clean and noisy sets. This paradigm, however, is prone to
significant degeneration under heavy label noise, as the number of clean
samples is too small for conventional methods to behave well. In this paper, we
introduce a novel framework, termed as LC-Booster, to explicitly tackle
learning under extreme noise. The core idea of LC-Booster is to incorporate
label correction into the sample selection, so that more purified samples,
through the reliable label correction, can be utilized for training, thereby
alleviating the confirmation bias. Experiments show that LC-Booster advances
state-of-the-art results on several noisy-label benchmarks, including CIFAR-10,
CIFAR-100, Clothing1M and WebVision. Remarkably, under the extreme 90\% noise
ratio, LC-Booster achieves 93.5\% and 48.4\% accuracy on CIFAR-10 and
CIFAR-100, surpassing the state-of-the-art by 1.6\% and 7.2\% respectively.
Related papers
- Addressing Long-Tail Noisy Label Learning Problems: a Two-Stage Solution
with Label Refurbishment Considering Label Rarity [13.490974408726323]
We introduce an effective two-stage approach by combining soft-label refurbishing with multi-expert ensemble learning.
In the first stage of robust soft label refurbishing, we acquire unbiased features through contrastive learning.
In the second stage, our label refurbishment method is applied to obtain soft labels for multi-expert ensemble learning.
arXiv Detail & Related papers (2024-03-04T08:06:57Z) - Class Prototype-based Cleaner for Label Noise Learning [73.007001454085]
Semi-supervised learning methods are current SOTA solutions to the noisy-label learning problem.
We propose a simple yet effective solution, named textbfClass textbfPrototype-based label noise textbfCleaner.
arXiv Detail & Related papers (2022-12-21T04:56:41Z) - Neighborhood Collective Estimation for Noisy Label Identification and
Correction [92.20697827784426]
Learning with noisy labels (LNL) aims at designing strategies to improve model performance and generalization by mitigating the effects of model overfitting to noisy labels.
Recent advances employ the predicted label distributions of individual samples to perform noise verification and noisy label correction, easily giving rise to confirmation bias.
We propose Neighborhood Collective Estimation, in which the predictive reliability of a candidate sample is re-estimated by contrasting it against its feature-space nearest neighbors.
arXiv Detail & Related papers (2022-08-05T14:47:22Z) - UNICON: Combating Label Noise Through Uniform Selection and Contrastive
Learning [89.56465237941013]
We propose UNICON, a simple yet effective sample selection method which is robust to high label noise.
We obtain an 11.4% improvement over the current state-of-the-art on CIFAR100 dataset with a 90% noise rate.
arXiv Detail & Related papers (2022-03-28T07:36:36Z) - PARS: Pseudo-Label Aware Robust Sample Selection for Learning with Noisy
Labels [5.758073912084364]
We propose PARS: Pseudo-Label Aware Robust Sample Selection.
PARS exploits all training samples using both the raw/noisy labels and estimated/refurbished pseudo-labels via self-training.
Results show that PARS significantly outperforms the state of the art on extensive studies on the noisy CIFAR-10 and CIFAR-100 datasets.
arXiv Detail & Related papers (2022-01-26T09:31:55Z) - Sample Prior Guided Robust Model Learning to Suppress Noisy Labels [8.119439844514973]
We propose PGDF, a novel framework to learn a deep model to suppress noise by generating the samples' prior knowledge.
Our framework can save more informative hard clean samples into the cleanly labeled set.
We evaluate our method using synthetic datasets based on CIFAR-10 and CIFAR-100, as well as on the real-world datasets WebVision and Clothing1M.
arXiv Detail & Related papers (2021-12-02T13:09:12Z) - Training Classifiers that are Universally Robust to All Label Noise
Levels [91.13870793906968]
Deep neural networks are prone to overfitting in the presence of label noise.
We propose a distillation-based framework that incorporates a new subcategory of Positive-Unlabeled learning.
Our framework generally outperforms at medium to high noise levels.
arXiv Detail & Related papers (2021-05-27T13:49:31Z) - LongReMix: Robust Learning with High Confidence Samples in a Noisy Label
Environment [33.376639002442914]
We propose the new 2-stage noisy-label training algorithm LongReMix.
We test LongReMix on the noisy-label benchmarks CIFAR-10, CIFAR-100, WebVision, Clothing1M, and Food101-N.
Our approach achieves state-of-the-art performance in most datasets.
arXiv Detail & Related papers (2021-03-06T18:48:40Z) - In Defense of Pseudo-Labeling: An Uncertainty-Aware Pseudo-label
Selection Framework for Semi-Supervised Learning [53.1047775185362]
Pseudo-labeling (PL) is a general SSL approach that does not have this constraint but performs relatively poorly in its original formulation.
We argue that PL underperforms due to the erroneous high confidence predictions from poorly calibrated models.
We propose an uncertainty-aware pseudo-label selection (UPS) framework which improves pseudo labeling accuracy by drastically reducing the amount of noise encountered in the training process.
arXiv Detail & Related papers (2021-01-15T23:29:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.