Extended T: Learning with Mixed Closed-set and Open-set Noisy Labels
- URL: http://arxiv.org/abs/2012.00932v1
- Date: Wed, 2 Dec 2020 02:42:45 GMT
- Title: Extended T: Learning with Mixed Closed-set and Open-set Noisy Labels
- Authors: Xiaobo Xia, Tongliang Liu, Bo Han, Nannan Wang, Jiankang Deng, Jiatong
Li, Yinian Mao
- Abstract summary: The label noise transition matrix $T$ reflects the probabilities that true labels flip into noisy ones.
In this paper, we focus on learning under the mixed closed-set and open-set label noise.
Our method can better model the mixed label noise, following its more robust performance than the prior state-of-the-art label-noise learning methods.
- Score: 86.5943044285146
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The label noise transition matrix $T$, reflecting the probabilities that true
labels flip into noisy ones, is of vital importance to model label noise and
design statistically consistent classifiers. The traditional transition matrix
is limited to model closed-set label noise, where noisy training data has true
class labels within the noisy label set. It is unfitted to employ such a
transition matrix to model open-set label noise, where some true class labels
are outside the noisy label set. Thus when considering a more realistic
situation, i.e., both closed-set and open-set label noise occurs, existing
methods will undesirably give biased solutions. Besides, the traditional
transition matrix is limited to model instance-independent label noise, which
may not perform well in practice. In this paper, we focus on learning under the
mixed closed-set and open-set label noise. We address the aforementioned issues
by extending the traditional transition matrix to be able to model mixed label
noise, and further to the cluster-dependent transition matrix to better
approximate the instance-dependent label noise in real-world applications. We
term the proposed transition matrix as the cluster-dependent extended
transition matrix. An unbiased estimator (i.e., extended $T$-estimator) has
been designed to estimate the cluster-dependent extended transition matrix by
only exploiting the noisy data. Comprehensive synthetic and real experiments
validate that our method can better model the mixed label noise, following its
more robust performance than the prior state-of-the-art label-noise learning
methods.
Related papers
- Extracting Clean and Balanced Subset for Noisy Long-tailed Classification [66.47809135771698]
We develop a novel pseudo labeling method using class prototypes from the perspective of distribution matching.
By setting a manually-specific probability measure, we can reduce the side-effects of noisy and long-tailed data simultaneously.
Our method can extract this class-balanced subset with clean labels, which brings effective performance gains for long-tailed classification with label noise.
arXiv Detail & Related papers (2024-04-10T07:34:37Z) - Multi-Label Noise Transition Matrix Estimation with Label Correlations:
Theory and Algorithm [73.94839250910977]
Noisy multi-label learning has garnered increasing attention due to the challenges posed by collecting large-scale accurate labels.
The introduction of transition matrices can help model multi-label noise and enable the development of statistically consistent algorithms.
We propose a novel estimator that leverages label correlations without the need for anchor points or precise fitting of noisy class posteriors.
arXiv Detail & Related papers (2023-09-22T08:35:38Z) - Rethinking the Value of Labels for Instance-Dependent Label Noise
Learning [43.481591776038144]
noisy labels in real-world applications often depend on both the true label and the features.
In this work, we tackle instance-dependent label noise with a novel deep generative model that avoids explicitly modeling the noise transition matrix.
Our algorithm leverages casual representation learning and simultaneously identifies the high-level content and style latent factors from the data.
arXiv Detail & Related papers (2023-05-10T15:29:07Z) - A Realistic Simulation Framework for Learning with Label Noise [17.14439597393087]
We show that this framework generates synthetic noisy labels that exhibit important characteristics of the label noise.
We also benchmark several existing algorithms for learning with noisy labels.
We propose a new technique, Label Quality Model (LQM), that leverages annotator features to predict and correct against noisy labels.
arXiv Detail & Related papers (2021-07-23T18:53:53Z) - Rethinking Noisy Label Models: Labeler-Dependent Noise with Adversarial
Awareness [2.1930130356902207]
We propose a principled model of label noise that generalizes instance-dependent noise to multiple labelers.
Under our labeler-dependent model, label noise manifests itself under two modalities: natural error of good-faith labelers, and adversarial labels provided by malicious actors.
We present two adversarial attack vectors that more accurately reflect the label noise that may be encountered in real-world settings.
arXiv Detail & Related papers (2021-05-28T19:58:18Z) - Learning Noise Transition Matrix from Only Noisy Labels via Total
Variation Regularization [88.91872713134342]
We propose a theoretically grounded method that can estimate the noise transition matrix and learn a classifier simultaneously.
We show the effectiveness of the proposed method through experiments on benchmark and real-world datasets.
arXiv Detail & Related papers (2021-02-04T05:09:18Z) - Meta Transition Adaptation for Robust Deep Learning with Noisy Labels [61.8970957519509]
This study proposes a new meta-transition-learning strategy for the task.
Specifically, through the sound guidance of a small set of meta data with clean labels, the noise transition matrix and the classifier parameters can be mutually ameliorated.
Our method can more accurately extract the transition matrix, naturally following its more robust performance than prior arts.
arXiv Detail & Related papers (2020-06-10T07:27:25Z) - Multi-Class Classification from Noisy-Similarity-Labeled Data [98.13491369929798]
We propose a method for learning from only noisy-similarity-labeled data.
We use a noise transition matrix to bridge the class-posterior probability between clean and noisy data.
We build a novel learning system which can assign noise-free class labels for instances.
arXiv Detail & Related papers (2020-02-16T05:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.