Noisy Labels Can Induce Good Representations
- URL: http://arxiv.org/abs/2012.12896v1
- Date: Wed, 23 Dec 2020 18:58:05 GMT
- Title: Noisy Labels Can Induce Good Representations
- Authors: Jingling Li, Mozhi Zhang, Keyulu Xu, John P. Dickerson, Jimmy Ba
- Abstract summary: We study how architecture affects learning with noisy labels.
We show that training with noisy labels can induce useful hidden representations, even when the model generalizes poorly.
This finding leads to a simple method to improve models trained on noisy labels.
- Score: 53.47668632785373
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The current success of deep learning depends on large-scale labeled datasets.
In practice, high-quality annotations are expensive to collect, but noisy
annotations are more affordable. Previous works report mixed empirical results
when training with noisy labels: neural networks can easily memorize random
labels, but they can also generalize from noisy labels. To explain this puzzle,
we study how architecture affects learning with noisy labels. We observe that
if an architecture "suits" the task, training with noisy labels can induce
useful hidden representations, even when the model generalizes poorly; i.e.,
the last few layers of the model are more negatively affected by noisy labels.
This finding leads to a simple method to improve models trained on noisy
labels: replacing the final dense layers with a linear model, whose weights are
learned from a small set of clean data. We empirically validate our findings
across three architectures (Convolutional Neural Networks, Graph Neural
Networks, and Multi-Layer Perceptrons) and two domains (graph algorithmic tasks
and image classification). Furthermore, we achieve state-of-the-art results on
image classification benchmarks by combining our method with existing
approaches on noisy label training.
Related papers
- ERASE: Error-Resilient Representation Learning on Graphs for Label Noise
Tolerance [53.73316938815873]
We propose a method called ERASE (Error-Resilient representation learning on graphs for lAbel noiSe tolerancE) to learn representations with error tolerance.
ERASE combines prototype pseudo-labels with propagated denoised labels and updates representations with error resilience.
Our method can outperform multiple baselines with clear margins in broad noise levels and enjoy great scalability.
arXiv Detail & Related papers (2023-12-13T17:59:07Z) - Learning to Aggregate and Refine Noisy Labels for Visual Sentiment
Analysis [69.48582264712854]
We propose a robust learning method to perform robust visual sentiment analysis.
Our method relies on an external memory to aggregate and filter noisy labels during training.
We establish a benchmark for visual sentiment analysis with label noise using publicly available datasets.
arXiv Detail & Related papers (2021-09-15T18:18:28Z) - A Survey on Deep Learning with Noisy Labels: How to train your model
when you cannot trust on the annotations? [21.562089974755125]
Several approaches have been proposed to improve the training of deep learning models in the presence of noisy labels.
This paper presents a survey on the main techniques in literature, in which we classify the algorithm in the following groups: robust losses, sample weighting, sample selection, meta-learning, and combined approaches.
arXiv Detail & Related papers (2020-12-05T15:45:20Z) - Grafit: Learning fine-grained image representations with coarse labels [114.17782143848315]
This paper tackles the problem of learning a finer representation than the one provided by training labels.
By jointly leveraging the coarse labels and the underlying fine-grained latent space, it significantly improves the accuracy of category-level retrieval methods.
arXiv Detail & Related papers (2020-11-25T19:06:26Z) - Combining Self-Supervised and Supervised Learning with Noisy Labels [41.627404715407586]
convolutional neural networks (CNNs) can easily overfit noisy labels.
It has been a great challenge to train CNNs against them robustly.
arXiv Detail & Related papers (2020-11-16T18:13:41Z) - Attention-Aware Noisy Label Learning for Image Classification [97.26664962498887]
Deep convolutional neural networks (CNNs) learned on large-scale labeled samples have achieved remarkable progress in computer vision.
The cheapest way to obtain a large body of labeled visual data is to crawl from websites with user-supplied labels, such as Flickr.
This paper proposes the attention-aware noisy label learning approach to improve the discriminative capability of the network trained on datasets with potential label noise.
arXiv Detail & Related papers (2020-09-30T15:45:36Z) - Label Noise Types and Their Effects on Deep Learning [0.0]
In this work, we provide a detailed analysis of the effects of different kinds of label noise on learning.
We propose a generic framework to generate feature-dependent label noise, which we show to be the most challenging case for learning.
For the ease of other researchers to test their algorithms with noisy labels, we share corrupted labels for the most commonly used benchmark datasets.
arXiv Detail & Related papers (2020-03-23T18:03:39Z) - Multi-Class Classification from Noisy-Similarity-Labeled Data [98.13491369929798]
We propose a method for learning from only noisy-similarity-labeled data.
We use a noise transition matrix to bridge the class-posterior probability between clean and noisy data.
We build a novel learning system which can assign noise-free class labels for instances.
arXiv Detail & Related papers (2020-02-16T05:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.