Beyond Class-Conditional Assumption: A Primary Attempt to Combat
Instance-Dependent Label Noise
- URL: http://arxiv.org/abs/2012.05458v1
- Date: Thu, 10 Dec 2020 05:16:18 GMT
- Title: Beyond Class-Conditional Assumption: A Primary Attempt to Combat
Instance-Dependent Label Noise
- Authors: Pengfei Chen, Junjie Ye, Guangyong Chen, Jingwei Zhao, Pheng-Ann Heng
- Abstract summary: Supervised learning under label noise has seen numerous advances recently.
We present a theoretical hypothesis testing and prove that noise in real-world dataset is unlikely to be CCN.
We formalize an algorithm to generate controllable instance-dependent noise (IDN)
- Score: 51.66448070984615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Supervised learning under label noise has seen numerous advances recently,
while existing theoretical findings and empirical results broadly build up on
the class-conditional noise (CCN) assumption that the noise is independent of
input features given the true label. In this work, we present a theoretical
hypothesis testing and prove that noise in real-world dataset is unlikely to be
CCN, which confirms that label noise should depend on the instance and
justifies the urgent need to go beyond the CCN assumption.The theoretical
results motivate us to study the more general and practical-relevant
instance-dependent noise (IDN). To stimulate the development of theory and
methodology on IDN, we formalize an algorithm to generate controllable IDN and
present both theoretical and empirical evidence to show that IDN is
semantically meaningful and challenging. As a primary attempt to combat IDN, we
present a tiny algorithm termed self-evolution average label (SEAL), which not
only stands out under IDN with various noise fractions, but also improves the
generalization on real-world noise benchmark Clothing1M. Our code is released.
Notably, our theoretical analysis in Section 2 provides rigorous motivations
for studying IDN, which is an important topic that deserves more research
attention in future.
Related papers
- Label Noise: Ignorance Is Bliss [20.341746708177055]
We establish a new theoretical framework for learning under multi-class, instance-dependent label noise.
Our findings support the simple emphNoise Ignorant Empirical Risk Minimization (NI-ERM) principle, which minimizes empirical risk while ignoring label noise.
arXiv Detail & Related papers (2024-10-31T17:03:25Z) - Understanding and Mitigating the Label Noise in Pre-training on
Downstream Tasks [91.15120211190519]
This paper aims to understand the nature of noise in pre-training datasets and to mitigate its impact on downstream tasks.
We propose a light-weight black-box tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise.
arXiv Detail & Related papers (2023-09-29T06:18:15Z) - Feature Noise Boosts DNN Generalization under Label Noise [65.36889005555669]
The presence of label noise in the training data has a profound impact on the generalization of deep neural networks (DNNs)
In this study, we introduce and theoretically demonstrate a simple feature noise method, which directly adds noise to the features of training data.
arXiv Detail & Related papers (2023-08-03T08:31:31Z) - Noise Injection Node Regularization for Robust Learning [0.0]
Noise Injection Node Regularization (NINR) is a method of injecting structured noise into Deep Neural Networks (DNN) during the training stage, resulting in an emergent regularizing effect.
We present theoretical and empirical evidence for substantial improvement in robustness against various test data perturbations for feed-forward DNNs when trained under NINR.
arXiv Detail & Related papers (2022-10-27T20:51:15Z) - Worst-Case Dynamic Power Distribution Network Noise Prediction Using
Convolutional Neural Network [14.144190519120167]
Worst-case dynamic PDN noise analysis is an essential step in PDN sign-off to ensure the performance and reliability of chips.
This paper proposes an efficient and scalable framework for the worst-case dynamic PDN noise prediction.
arXiv Detail & Related papers (2022-04-27T08:37:10Z) - The Optimal Noise in Noise-Contrastive Learning Is Not What You Think [80.07065346699005]
We show that deviating from this assumption can actually lead to better statistical estimators.
In particular, the optimal noise distribution is different from the data's and even from a different family.
arXiv Detail & Related papers (2022-03-02T13:59:20Z) - Learning with Noisy Labels Revisited: A Study Using Real-World Human
Annotations [54.400167806154535]
Existing research on learning with noisy labels mainly focuses on synthetic label noise.
This work presents two new benchmark datasets (CIFAR-10N, CIFAR-100N)
We show that real-world noisy labels follow an instance-dependent pattern rather than the classically adopted class-dependent ones.
arXiv Detail & Related papers (2021-10-22T22:42:11Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - Unsupervised Neural Universal Denoiser for Finite-Input General-Output
Noisy Channel [25.26787589154647]
We devise a novel neural network-based universal denoiser for the finite-input, general-output (FIGO) channel.
Based on the assumption of known noisy channel densities, we train the network such that it can denoise as well as the best sliding window denoiser for any given underlying clean source data.
arXiv Detail & Related papers (2020-03-05T17:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.