Multi-Class Classification from Noisy-Similarity-Labeled Data
- URL: http://arxiv.org/abs/2002.06508v1
- Date: Sun, 16 Feb 2020 05:10:21 GMT
- Title: Multi-Class Classification from Noisy-Similarity-Labeled Data
- Authors: Songhua Wu, Xiaobo Xia, Tongliang Liu, Bo Han, Mingming Gong, Nannan
Wang, Haifeng Liu, Gang Niu
- Abstract summary: We propose a method for learning from only noisy-similarity-labeled data.
We use a noise transition matrix to bridge the class-posterior probability between clean and noisy data.
We build a novel learning system which can assign noise-free class labels for instances.
- Score: 98.13491369929798
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A similarity label indicates whether two instances belong to the same class
while a class label shows the class of the instance. Without class labels, a
multi-class classifier could be learned from similarity-labeled pairwise data
by meta classification learning. However, since the similarity label is less
informative than the class label, it is more likely to be noisy. Deep neural
networks can easily remember noisy data, leading to overfitting in
classification. In this paper, we propose a method for learning from only
noisy-similarity-labeled data. Specifically, to model the noise, we employ a
noise transition matrix to bridge the class-posterior probability between clean
and noisy data. We further estimate the transition matrix from only noisy data
and build a novel learning system to learn a classifier which can assign
noise-free class labels for instances. Moreover, we theoretically justify how
our proposed method generalizes for learning classifiers. Experimental results
demonstrate the superiority of the proposed method over the state-of-the-art
method on benchmark-simulated and real-world noisy-label datasets.
Related papers
- Estimating Noisy Class Posterior with Part-level Labels for Noisy Label Learning [13.502549812291878]
Existing methods typically learn noisy class posteriors by training a classification model with noisy labels.
This paper proposes to augment the supervised information with part-level labels, encouraging the model to focus on and integrate richer information from various parts.
Our method is theoretically sound, while experiments show that it is empirically effective in synthetic and real-world noisy benchmarks.
arXiv Detail & Related papers (2024-05-08T12:13:40Z) - Multi-Label Noise Transition Matrix Estimation with Label Correlations:
Theory and Algorithm [73.94839250910977]
Noisy multi-label learning has garnered increasing attention due to the challenges posed by collecting large-scale accurate labels.
The introduction of transition matrices can help model multi-label noise and enable the development of statistically consistent algorithms.
We propose a novel estimator that leverages label correlations without the need for anchor points or precise fitting of noisy class posteriors.
arXiv Detail & Related papers (2023-09-22T08:35:38Z) - Instance-dependent Label-noise Learning under a Structural Causal Model [92.76400590283448]
Label noise will degenerate the performance of deep learning algorithms.
By leveraging a structural causal model, we propose a novel generative approach for instance-dependent label-noise learning.
arXiv Detail & Related papers (2021-09-07T10:42:54Z) - Learning from Multiple Annotators by Incorporating Instance Features [15.643325526074804]
Learning from multiple annotators aims to induce a high-quality classifier from training instances.
Most existing methods adopt class-level confusion matrices of annotators that observed labels do not depend on the instance features.
We propose a noise transition matrix, which incorporates the influence of instance features on annotators' performance based on confusion matrices.
arXiv Detail & Related papers (2021-06-29T08:07:24Z) - Noisy Labels Can Induce Good Representations [53.47668632785373]
We study how architecture affects learning with noisy labels.
We show that training with noisy labels can induce useful hidden representations, even when the model generalizes poorly.
This finding leads to a simple method to improve models trained on noisy labels.
arXiv Detail & Related papers (2020-12-23T18:58:05Z) - Extended T: Learning with Mixed Closed-set and Open-set Noisy Labels [86.5943044285146]
The label noise transition matrix $T$ reflects the probabilities that true labels flip into noisy ones.
In this paper, we focus on learning under the mixed closed-set and open-set label noise.
Our method can better model the mixed label noise, following its more robust performance than the prior state-of-the-art label-noise learning methods.
arXiv Detail & Related papers (2020-12-02T02:42:45Z) - Error-Bounded Correction of Noisy Labels [17.510654621245656]
We show that the prediction of a noisy classifier can indeed be a good indicator of whether the label of a training data is clean.
Based on the theoretical result, we propose a novel algorithm that corrects the labels based on the noisy classifier prediction.
We incorporate our label correction algorithm into the training of deep neural networks and train models that achieve superior testing performance on multiple public datasets.
arXiv Detail & Related papers (2020-11-19T19:23:23Z) - Class2Simi: A Noise Reduction Perspective on Learning with Noisy Labels [98.13491369929798]
We propose a framework called Class2Simi, which transforms data points with noisy class labels to data pairs with noisy similarity labels.
Class2Simi is computationally efficient because not only this transformation is on-the-fly in mini-batches, but also it just changes loss on top of model prediction into a pairwise manner.
arXiv Detail & Related papers (2020-06-14T07:55:32Z) - Label Noise Types and Their Effects on Deep Learning [0.0]
In this work, we provide a detailed analysis of the effects of different kinds of label noise on learning.
We propose a generic framework to generate feature-dependent label noise, which we show to be the most challenging case for learning.
For the ease of other researchers to test their algorithms with noisy labels, we share corrupted labels for the most commonly used benchmark datasets.
arXiv Detail & Related papers (2020-03-23T18:03:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.