Feature Noise Boosts DNN Generalization under Label Noise
- URL: http://arxiv.org/abs/2308.01609v1
- Date: Thu, 3 Aug 2023 08:31:31 GMT
- Title: Feature Noise Boosts DNN Generalization under Label Noise
- Authors: Lu Zeng, Xuan Chen, Xiaoshuang Shi, Heng Tao Shen
- Abstract summary: The presence of label noise in the training data has a profound impact on the generalization of deep neural networks (DNNs)
In this study, we introduce and theoretically demonstrate a simple feature noise method, which directly adds noise to the features of training data.
- Score: 65.36889005555669
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The presence of label noise in the training data has a profound impact on the
generalization of deep neural networks (DNNs). In this study, we introduce and
theoretically demonstrate a simple feature noise method, which directly adds
noise to the features of training data, can enhance the generalization of DNNs
under label noise. Specifically, we conduct theoretical analyses to reveal that
label noise leads to weakened DNN generalization by loosening the PAC-Bayes
generalization bound, and feature noise results in better DNN generalization by
imposing an upper bound on the mutual information between the model weights and
the features, which constrains the PAC-Bayes generalization bound. Furthermore,
to ensure effective generalization of DNNs in the presence of label noise, we
conduct application analyses to identify the optimal types and levels of
feature noise to add for obtaining desirable label noise generalization.
Finally, extensive experimental results on several popular datasets demonstrate
the feature noise method can significantly enhance the label noise
generalization of the state-of-the-art label noise method.
Related papers
- SoftPatch: Unsupervised Anomaly Detection with Noisy Data [67.38948127630644]
This paper considers label-level noise in image sensory anomaly detection for the first time.
We propose a memory-based unsupervised AD method, SoftPatch, which efficiently denoises the data at the patch level.
Compared with existing methods, SoftPatch maintains a strong modeling ability of normal data and alleviates the overconfidence problem in coreset.
arXiv Detail & Related papers (2024-03-21T08:49:34Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - NoisywikiHow: A Benchmark for Learning with Real-world Noisy Labels in
Natural Language Processing [26.678589684142548]
Large-scale datasets in the real world inevitably involve label noise.
Deep models can gradually overfit noisy labels and thus degrade generalization performance.
To mitigate the effects of label noise, learning with noisy labels (LNL) methods are designed to achieve better generalization performance.
arXiv Detail & Related papers (2023-05-18T05:01:04Z) - Latent Class-Conditional Noise Model [54.56899309997246]
We introduce a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework.
We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels.
Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples.
arXiv Detail & Related papers (2023-02-19T15:24:37Z) - Noise Injection Node Regularization for Robust Learning [0.0]
Noise Injection Node Regularization (NINR) is a method of injecting structured noise into Deep Neural Networks (DNN) during the training stage, resulting in an emergent regularizing effect.
We present theoretical and empirical evidence for substantial improvement in robustness against various test data perturbations for feed-forward DNNs when trained under NINR.
arXiv Detail & Related papers (2022-10-27T20:51:15Z) - Open-set Label Noise Can Improve Robustness Against Inherent Label Noise [27.885927200376386]
We show that open-set noisy labels can be non-toxic and even benefit the robustness against inherent noisy labels.
We propose a simple yet effective regularization by introducing Open-set samples with Dynamic Noisy Labels (ODNL) into training.
arXiv Detail & Related papers (2021-06-21T07:15:50Z) - Analysing the Noise Model Error for Realistic Noisy Label Data [14.766574408868806]
We study the quality of estimated noise models from the theoretical side by deriving the expected error of the noise model.
We also publish NoisyNER, a new noisy label dataset from the NLP domain.
arXiv Detail & Related papers (2021-01-24T17:45:15Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - Beyond Class-Conditional Assumption: A Primary Attempt to Combat
Instance-Dependent Label Noise [51.66448070984615]
Supervised learning under label noise has seen numerous advances recently.
We present a theoretical hypothesis testing and prove that noise in real-world dataset is unlikely to be CCN.
We formalize an algorithm to generate controllable instance-dependent noise (IDN)
arXiv Detail & Related papers (2020-12-10T05:16:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.