GANs for learning from very high class conditional noisy labels
- URL: http://arxiv.org/abs/2010.09577v1
- Date: Mon, 19 Oct 2020 15:01:11 GMT
- Title: GANs for learning from very high class conditional noisy labels
- Authors: Sandhya Tripathi and N Hemachandra
- Abstract summary: We use Generative Adversarial Networks (GANs) to design a class conditional label noise (CCN) robust scheme for binary classification.
It first generates a set of correctly labelled data points from noisy labelled data and 0.1% or 1% clean labels.
- Score: 1.6516902135723865
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We use Generative Adversarial Networks (GANs) to design a class conditional
label noise (CCN) robust scheme for binary classification. It first generates a
set of correctly labelled data points from noisy labelled data and 0.1% or 1%
clean labels such that the generated and true (clean) labelled data
distributions are close; generated labelled data is used to learn a good
classifier. The mode collapse problem while generating correct feature-label
pairs and the problem of skewed feature-label dimension ratio ($\sim$ 784:1)
are avoided by using Wasserstein GAN and a simple data representation change.
Another WGAN with information-theoretic flavour on top of the new
representation is also proposed. The major advantage of both schemes is their
significant improvement over the existing ones in presence of very high CCN
rates, without either estimating or cross-validating over the noise rates. We
proved that KL divergence between clean and noisy distribution increases w.r.t.
noise rates in symmetric label noise model; can be extended to high CCN rates.
This implies that our schemes perform well due to the adversarial nature of
GANs. Further, use of generative approach (learning clean joint distribution)
while handling noise enables our schemes to perform better than discriminative
approaches like GLC, LDMI and GCE; even when the classes are highly imbalanced.
Using Friedman F test and Nemenyi posthoc test, we showed that on high
dimensional binary class synthetic, MNIST and Fashion MNIST datasets, our
schemes outperform the existing methods and demonstrate consistent performance
across noise rates.
Related papers
- Extracting Clean and Balanced Subset for Noisy Long-tailed Classification [66.47809135771698]
We develop a novel pseudo labeling method using class prototypes from the perspective of distribution matching.
By setting a manually-specific probability measure, we can reduce the side-effects of noisy and long-tailed data simultaneously.
Our method can extract this class-balanced subset with clean labels, which brings effective performance gains for long-tailed classification with label noise.
arXiv Detail & Related papers (2024-04-10T07:34:37Z) - CSGNN: Conquering Noisy Node labels via Dynamic Class-wise Selection [45.83801634434111]
We introduce a novel Class-wise Selection for Graph Neural Networks, dubbed CSGNN.
To tackle the class imbalance issue, we introduce a dynamic class-wise selection mechanism, leveraging the clustering technique to identify clean nodes.
To alleviate the problem of noisy labels, built on the concept of the memorization effect, CSGNN prioritizes learning from clean nodes before noisy ones.
arXiv Detail & Related papers (2023-11-20T00:57:30Z) - Latent Class-Conditional Noise Model [54.56899309997246]
We introduce a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework.
We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels.
Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples.
arXiv Detail & Related papers (2023-02-19T15:24:37Z) - Learning from Noisy Labels with Coarse-to-Fine Sample Credibility
Modeling [22.62790706276081]
Training deep neural network (DNN) with noisy labels is practically challenging.
Previous efforts tend to handle part or full data in a unified denoising flow.
We propose a coarse-to-fine robust learning method called CREMA to handle noisy data in a divide-and-conquer manner.
arXiv Detail & Related papers (2022-08-23T02:06:38Z) - Neighborhood Collective Estimation for Noisy Label Identification and
Correction [92.20697827784426]
Learning with noisy labels (LNL) aims at designing strategies to improve model performance and generalization by mitigating the effects of model overfitting to noisy labels.
Recent advances employ the predicted label distributions of individual samples to perform noise verification and noisy label correction, easily giving rise to confirmation bias.
We propose Neighborhood Collective Estimation, in which the predictive reliability of a candidate sample is re-estimated by contrasting it against its feature-space nearest neighbors.
arXiv Detail & Related papers (2022-08-05T14:47:22Z) - Training Classifiers that are Universally Robust to All Label Noise
Levels [91.13870793906968]
Deep neural networks are prone to overfitting in the presence of label noise.
We propose a distillation-based framework that incorporates a new subcategory of Positive-Unlabeled learning.
Our framework generally outperforms at medium to high noise levels.
arXiv Detail & Related papers (2021-05-27T13:49:31Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - A Second-Order Approach to Learning with Instance-Dependent Label Noise [58.555527517928596]
The presence of label noise often misleads the training of deep neural networks.
We show that the errors in human-annotated labels are more likely to be dependent on the difficulty levels of tasks.
arXiv Detail & Related papers (2020-12-22T06:36:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.