Approximating Instance-Dependent Noise via Instance-Confidence Embedding
- URL: http://arxiv.org/abs/2103.13569v1
- Date: Thu, 25 Mar 2021 02:33:30 GMT
- Title: Approximating Instance-Dependent Noise via Instance-Confidence Embedding
- Authors: Yivan Zhang, Masashi Sugiyama
- Abstract summary: Label noise in multiclass classification is a major obstacle to the deployment of learning systems.
We investigate the instance-dependent noise (IDN) model and propose an efficient approximation of IDN to capture the instance-specific label corruption.
- Score: 87.65718705642819
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Label noise in multiclass classification is a major obstacle to the
deployment of learning systems. However, unlike the widely used
class-conditional noise (CCN) assumption that the noisy label is independent of
the input feature given the true label, label noise in real-world datasets can
be aleatory and heavily dependent on individual instances. In this work, we
investigate the instance-dependent noise (IDN) model and propose an efficient
approximation of IDN to capture the instance-specific label corruption.
Concretely, noting the fact that most columns of the IDN transition matrix have
only limited influence on the class-posterior estimation, we propose a
variational approximation that uses a single-scalar confidence parameter. To
cope with the situation where the mapping from the instance to its confidence
value could vary significantly for two adjacent instances, we suggest using
instance embedding that assigns a trainable parameter to each instance. The
resulting instance-confidence embedding (ICE) method not only performs well
under label noise but also can effectively detect ambiguous or mislabeled
instances. We validate its utility on various image and text classification
tasks.
Related papers
- InstanT: Semi-supervised Learning with Instance-dependent Thresholds [75.91684890150283]
We propose the study of instance-dependent thresholds, which has the highest degree of freedom compared with existing methods.
We devise a novel instance-dependent threshold function for all unlabeled instances by utilizing their instance-level ambiguity and the instance-dependent error rates of pseudo-labels.
arXiv Detail & Related papers (2023-10-29T05:31:43Z) - Instance-Dependent Noisy Label Learning via Graphical Modelling [30.922188228545906]
Noisy labels are troublesome in the ecosystem of deep learning because models can easily overfit them.
We present a new graphical modelling approach called InstanceGM that combines discriminative and generative models.
arXiv Detail & Related papers (2022-09-02T09:27:37Z) - S3: Supervised Self-supervised Learning under Label Noise [53.02249460567745]
In this paper we address the problem of classification in the presence of label noise.
In the heart of our method is a sample selection mechanism that relies on the consistency between the annotated label of a sample and the distribution of the labels in its neighborhood in the feature space.
Our method significantly surpasses previous methods on both CIFARCIFAR100 with artificial noise and real-world noisy datasets such as WebVision and ANIMAL-10N.
arXiv Detail & Related papers (2021-11-22T15:49:20Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - A Second-Order Approach to Learning with Instance-Dependent Label Noise [58.555527517928596]
The presence of label noise often misleads the training of deep neural networks.
We show that the errors in human-annotated labels are more likely to be dependent on the difficulty levels of tasks.
arXiv Detail & Related papers (2020-12-22T06:36:58Z) - Confidence Scores Make Instance-dependent Label-noise Learning Possible [129.84497190791103]
In learning with noisy labels, for every instance, its label can randomly walk to other classes following a transition distribution which is named a noise model.
We introduce confidence-scored instance-dependent noise (CSIDN), where each instance-label pair is equipped with a confidence score.
We find with the help of confidence scores, the transition distribution of each instance can be approximately estimated.
arXiv Detail & Related papers (2020-01-11T16:15:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.