MILD: Modeling the Instance Learning Dynamics for Learning with Noisy
Labels
- URL: http://arxiv.org/abs/2306.11560v2
- Date: Tue, 30 Jan 2024 12:55:08 GMT
- Title: MILD: Modeling the Instance Learning Dynamics for Learning with Noisy
Labels
- Authors: Chuanyang Hu, Shipeng Yan, Zhitong Gao, Xuming He
- Abstract summary: We propose an iterative selection approach based on the Weibull mixture model to identify clean data.
In particular, we measure the difficulty of memorization and memorize for each instance via the transition times between being misclassified and being memorized.
Our strategy outperforms existing noisy-label learning methods.
- Score: 19.650299232829546
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite deep learning has achieved great success, it often relies on a large
amount of training data with accurate labels, which are expensive and
time-consuming to collect. A prominent direction to reduce the cost is to learn
with noisy labels, which are ubiquitous in the real-world applications. A
critical challenge for such a learning task is to reduce the effect of network
memorization on the falsely-labeled data. In this work, we propose an iterative
selection approach based on the Weibull mixture model, which identifies clean
data by considering the overall learning dynamics of each data instance. In
contrast to the previous small-loss heuristics, we leverage the observation
that deep network is easy to memorize and hard to forget clean data. In
particular, we measure the difficulty of memorization and forgetting for each
instance via the transition times between being misclassified and being
memorized in training, and integrate them into a novel metric for selection.
Based on the proposed metric, we retain a subset of identified clean data and
repeat the selection procedure to iteratively refine the clean subset, which is
finally used for model training. To validate our method, we perform extensive
experiments on synthetic noisy datasets and real-world web data, and our
strategy outperforms existing noisy-label learning methods.
Related papers
- Stochastic Amortization: A Unified Approach to Accelerate Feature and Data Attribution [62.71425232332837]
We show that training amortized models with noisy labels is inexpensive and surprisingly effective.
This approach significantly accelerates several feature attribution and data valuation methods, often yielding an order of magnitude speedup over existing approaches.
arXiv Detail & Related papers (2024-01-29T03:42:37Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Combating Label Noise With A General Surrogate Model For Sample
Selection [84.61367781175984]
We propose to leverage the vision-language surrogate model CLIP to filter noisy samples automatically.
We validate the effectiveness of our proposed method on both real-world and synthetic noisy datasets.
arXiv Detail & Related papers (2023-10-16T14:43:27Z) - Late Stopping: Avoiding Confidently Learning from Mislabeled Examples [61.00103151680946]
We propose a new framework, Late Stopping, which leverages the intrinsic robust learning ability of DNNs through a prolonged training process.
We empirically observe that mislabeled and clean examples exhibit differences in the number of epochs required for them to be consistently and correctly classified.
Experimental results on benchmark-simulated and real-world noisy datasets demonstrate that the proposed method outperforms state-of-the-art counterparts.
arXiv Detail & Related papers (2023-08-26T12:43:25Z) - Leveraging Unlabeled Data to Track Memorization [15.4909376515404]
We propose a metric, called susceptibility, to gauge memorization for neural networks.
We empirically show the effectiveness of our metric in tracking memorization on various architectures and datasets.
arXiv Detail & Related papers (2022-12-08T18:36:41Z) - Learning with Neighbor Consistency for Noisy Labels [69.83857578836769]
We present a method for learning from noisy labels that leverages similarities between training examples in feature space.
We evaluate our method on datasets evaluating both synthetic (CIFAR-10, CIFAR-100) and realistic (mini-WebVision, Clothing1M, mini-ImageNet-Red) noise.
arXiv Detail & Related papers (2022-02-04T15:46:27Z) - Understanding Memorization from the Perspective of Optimization via
Efficient Influence Estimation [54.899751055620904]
We study the phenomenon of memorization with turn-over dropout, an efficient method to estimate influence and memorization, for data with true labels (real data) and data with random labels (random data)
Our main findings are: (i) For both real data and random data, the optimization of easy examples (e.g., real data) and difficult examples (e.g., random data) are conducted by the network simultaneously, with easy ones at a higher speed; (ii) For real data, a correct difficult example in the training dataset is more informative than an easy one.
arXiv Detail & Related papers (2021-12-16T11:34:23Z) - INN: A Method Identifying Clean-annotated Samples via Consistency Effect
in Deep Neural Networks [1.1470070927586016]
We introduce a new method called INN to refine clean labeled data from training data with noisy labels.
The INN method requires more computation but is much stable and powerful than the small-loss strategy.
arXiv Detail & Related papers (2021-06-29T09:06:21Z) - Learning from Noisy Labels for Entity-Centric Information Extraction [17.50856935207308]
We propose a simple co-regularization framework for entity-centric information extraction.
These models are jointly optimized with task-specific loss, and are regularized to generate similar predictions.
In the end, we can take any of the trained models for inference.
arXiv Detail & Related papers (2021-04-17T22:49:12Z) - Self-semi-supervised Learning to Learn from NoisyLabeled Data [3.18577806302116]
It is costly to obtain high-quality human-labeled data, leading to the active research area of training models robust to noisy labels.
In this project, we designed methods to more accurately differentiate clean and noisy labels and borrowed the wisdom of self-semi-supervised learning to train noisy labeled data.
arXiv Detail & Related papers (2020-11-03T02:31:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.