Investigating Why Contrastive Learning Benefits Robustness Against Label
Noise
- URL: http://arxiv.org/abs/2201.12498v1
- Date: Sat, 29 Jan 2022 05:19:26 GMT
- Title: Investigating Why Contrastive Learning Benefits Robustness Against Label
Noise
- Authors: Yihao Xue, Kyle Whitecross, Baharan Mirzasoleiman
- Abstract summary: Self-supervised contrastive learning has been shown to be very effective in preventing deep networks from overfitting noisy labels.
We rigorously prove that the representation matrix learned by contrastive learning boosts robustness.
- Score: 6.855361451300868
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Self-supervised contrastive learning has recently been shown to be very
effective in preventing deep networks from overfitting noisy labels. Despite
its empirical success, the theoretical understanding of the effect of
contrastive learning on boosting robustness is very limited. In this work, we
rigorously prove that the representation matrix learned by contrastive learning
boosts robustness, by having: (i) one prominent singular value corresponding to
every sub-class in the data, and remaining significantly smaller singular
values; and (ii) a large alignment between the prominent singular vector and
the clean labels of each sub-class. The above properties allow a linear layer
trained on the representations to quickly learn the clean labels, and prevent
it from overfitting the noise for a large number of training iterations. We
further show that the low-rank structure of the Jacobian of deep networks
pre-trained with contrastive learning allows them to achieve a superior
performance initially, when fine-tuned on noisy labels. Finally, we demonstrate
that the initial robustness provided by contrastive learning enables robust
training methods to achieve state-of-the-art performance under extreme noise
levels, e.g., an average of 27.18\% and 15.58\% increase in accuracy on
CIFAR-10 and CIFAR-100 with 80\% symmetric noisy labels, and 4.11\% increase in
accuracy on WebVision.
Related papers
- ERASE: Error-Resilient Representation Learning on Graphs for Label Noise
Tolerance [53.73316938815873]
We propose a method called ERASE (Error-Resilient representation learning on graphs for lAbel noiSe tolerancE) to learn representations with error tolerance.
ERASE combines prototype pseudo-labels with propagated denoised labels and updates representations with error resilience.
Our method can outperform multiple baselines with clear margins in broad noise levels and enjoy great scalability.
arXiv Detail & Related papers (2023-12-13T17:59:07Z) - Towards Harnessing Feature Embedding for Robust Learning with Noisy
Labels [44.133307197696446]
The memorization effect of deep neural networks (DNNs) plays a pivotal role in recent label noise learning methods.
We propose a novel feature embedding-based method for deep learning with label noise, termed LabEl NoiseDilution (LEND)
arXiv Detail & Related papers (2022-06-27T02:45:09Z) - Transductive CLIP with Class-Conditional Contrastive Learning [68.51078382124331]
We propose Transductive CLIP, a novel framework for learning a classification network with noisy labels from scratch.
A class-conditional contrastive learning mechanism is proposed to mitigate the reliance on pseudo labels.
ensemble labels is adopted as a pseudo label updating strategy to stabilize the training of deep neural networks with noisy labels.
arXiv Detail & Related papers (2022-06-13T14:04:57Z) - Meta Self-Refinement for Robust Learning with Weak Supervision [29.80743717767389]
We propose Meta Self-Refinement (MSR) to combat label noise from weak supervision.
MSR is robust against label noise in all settings and outperforms state-of-the-art methods by up to 11.4% in accuracy and 9.26% in F1 score.
arXiv Detail & Related papers (2022-05-15T14:04:30Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - Contrastive Learning Improves Model Robustness Under Label Noise [3.756550107432323]
We show that by initializing supervised robust methods using representations learned through contrastive learning leads to significantly improved performance under label noise.
Even the simplest method can outperform the state-of-the-art SSL method by more than 50% under high label noise when with contrastive learning.
arXiv Detail & Related papers (2021-04-19T00:27:58Z) - In Defense of Pseudo-Labeling: An Uncertainty-Aware Pseudo-label
Selection Framework for Semi-Supervised Learning [53.1047775185362]
Pseudo-labeling (PL) is a general SSL approach that does not have this constraint but performs relatively poorly in its original formulation.
We argue that PL underperforms due to the erroneous high confidence predictions from poorly calibrated models.
We propose an uncertainty-aware pseudo-label selection (UPS) framework which improves pseudo labeling accuracy by drastically reducing the amount of noise encountered in the training process.
arXiv Detail & Related papers (2021-01-15T23:29:57Z) - Coresets for Robust Training of Neural Networks against Noisy Labels [78.03027938765746]
We propose a novel approach with strong theoretical guarantees for robust training of deep networks trained with noisy labels.
We select weighted subsets (coresets) of clean data points that provide an approximately low-rank Jacobian matrix.
Our experiments corroborate our theory and demonstrate that deep networks trained on our subsets achieve a significantly superior performance compared to state-of-the art.
arXiv Detail & Related papers (2020-11-15T04:58:11Z) - Adversarial Self-Supervised Contrastive Learning [62.17538130778111]
Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions.
We propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples.
We present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data.
arXiv Detail & Related papers (2020-06-13T08:24:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.