Class2Simi: A Noise Reduction Perspective on Learning with Noisy Labels
- URL: http://arxiv.org/abs/2006.07831v2
- Date: Thu, 17 Jun 2021 10:10:10 GMT
- Title: Class2Simi: A Noise Reduction Perspective on Learning with Noisy Labels
- Authors: Songhua Wu, Xiaobo Xia, Tongliang Liu, Bo Han, Mingming Gong, Nannan
Wang, Haifeng Liu, Gang Niu
- Abstract summary: We propose a framework called Class2Simi, which transforms data points with noisy class labels to data pairs with noisy similarity labels.
Class2Simi is computationally efficient because not only this transformation is on-the-fly in mini-batches, but also it just changes loss on top of model prediction into a pairwise manner.
- Score: 98.13491369929798
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning with noisy labels has attracted a lot of attention in recent years,
where the mainstream approaches are in pointwise manners. Meanwhile, pairwise
manners have shown great potential in supervised metric learning and
unsupervised contrastive learning. Thus, a natural question is raised: does
learning in a pairwise manner mitigate label noise? To give an affirmative
answer, in this paper, we propose a framework called Class2Simi: it transforms
data points with noisy class labels to data pairs with noisy similarity labels,
where a similarity label denotes whether a pair shares the class label or not.
Through this transformation, the reduction of the noise rate is theoretically
guaranteed, and hence it is in principle easier to handle noisy similarity
labels. Amazingly, DNNs that predict the clean class labels can be trained from
noisy data pairs if they are first pretrained from noisy data points.
Class2Simi is computationally efficient because not only this transformation is
on-the-fly in mini-batches, but also it just changes loss computation on top of
model prediction into a pairwise manner. Its effectiveness is verified by
extensive experiments.
Related papers
- Extracting Clean and Balanced Subset for Noisy Long-tailed Classification [66.47809135771698]
We develop a novel pseudo labeling method using class prototypes from the perspective of distribution matching.
By setting a manually-specific probability measure, we can reduce the side-effects of noisy and long-tailed data simultaneously.
Our method can extract this class-balanced subset with clean labels, which brings effective performance gains for long-tailed classification with label noise.
arXiv Detail & Related papers (2024-04-10T07:34:37Z) - Multi-Label Noise Transition Matrix Estimation with Label Correlations:
Theory and Algorithm [73.94839250910977]
Noisy multi-label learning has garnered increasing attention due to the challenges posed by collecting large-scale accurate labels.
The introduction of transition matrices can help model multi-label noise and enable the development of statistically consistent algorithms.
We propose a novel estimator that leverages label correlations without the need for anchor points or precise fitting of noisy class posteriors.
arXiv Detail & Related papers (2023-09-22T08:35:38Z) - A Second-Order Approach to Learning with Instance-Dependent Label Noise [58.555527517928596]
The presence of label noise often misleads the training of deep neural networks.
We show that the errors in human-annotated labels are more likely to be dependent on the difficulty levels of tasks.
arXiv Detail & Related papers (2020-12-22T06:36:58Z) - Extended T: Learning with Mixed Closed-set and Open-set Noisy Labels [86.5943044285146]
The label noise transition matrix $T$ reflects the probabilities that true labels flip into noisy ones.
In this paper, we focus on learning under the mixed closed-set and open-set label noise.
Our method can better model the mixed label noise, following its more robust performance than the prior state-of-the-art label-noise learning methods.
arXiv Detail & Related papers (2020-12-02T02:42:45Z) - Error-Bounded Correction of Noisy Labels [17.510654621245656]
We show that the prediction of a noisy classifier can indeed be a good indicator of whether the label of a training data is clean.
Based on the theoretical result, we propose a novel algorithm that corrects the labels based on the noisy classifier prediction.
We incorporate our label correction algorithm into the training of deep neural networks and train models that achieve superior testing performance on multiple public datasets.
arXiv Detail & Related papers (2020-11-19T19:23:23Z) - EvidentialMix: Learning with Combined Open-set and Closed-set Noisy
Labels [30.268962418683955]
We study a new variant of the noisy label problem that combines the open-set and closed-set noisy labels.
Our results show that our method produces superior classification results and better feature representations than previous state-of-the-art methods.
arXiv Detail & Related papers (2020-11-11T11:15:32Z) - Multi-Class Classification from Noisy-Similarity-Labeled Data [98.13491369929798]
We propose a method for learning from only noisy-similarity-labeled data.
We use a noise transition matrix to bridge the class-posterior probability between clean and noisy data.
We build a novel learning system which can assign noise-free class labels for instances.
arXiv Detail & Related papers (2020-02-16T05:10:21Z) - Particle Competition and Cooperation for Semi-Supervised Learning with
Label Noise [6.247917165799351]
A graph-based semi-supervised learning approach based on Particle competition and cooperation was developed.
This paper presents a new particle competition and cooperation algorithm, specifically designed to increase the robustness to the presence of label noise.
It performs classification of unlabeled nodes and reclassification of the nodes affected by label noise in a unique process.
arXiv Detail & Related papers (2020-02-12T19:44:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.