Learning to Clean: Reinforcement Learning for Noisy Label Correction
- URL: http://arxiv.org/abs/2511.19808v1
- Date: Tue, 25 Nov 2025 00:32:03 GMT
- Title: Learning to Clean: Reinforcement Learning for Noisy Label Correction
- Authors: Marzi Heidari, Hanping Zhang, Yuhong Guo,
- Abstract summary: This paper introduces a novel framework that conceptualizes noisy label correction as a reinforcement learning problem.<n>The proposed approach, Reinforcement Learning for Noisy Label Correction (RLNLC), defines a comprehensive state space representing data and their associated labels.<n>The effectiveness of RLNLC is demonstrated through extensive experiments on multiple benchmark datasets.
- Score: 25.643437301724365
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The challenge of learning with noisy labels is significant in machine learning, as it can severely degrade the performance of prediction models if not addressed properly. This paper introduces a novel framework that conceptualizes noisy label correction as a reinforcement learning (RL) problem. The proposed approach, Reinforcement Learning for Noisy Label Correction (RLNLC), defines a comprehensive state space representing data and their associated labels, an action space that indicates possible label corrections, and a reward mechanism that evaluates the efficacy of label corrections. RLNLC learns a deep feature representation based policy network to perform label correction through reinforcement learning, utilizing an actor-critic method. The learned policy is subsequently deployed to iteratively correct noisy training labels and facilitate the training of the prediction model. The effectiveness of RLNLC is demonstrated through extensive experiments on multiple benchmark datasets, where it consistently outperforms existing state-of-the-art techniques for learning with noisy labels.
Related papers
- Bridging Weakly-Supervised Learning and VLM Distillation: Noisy Partial Label Learning for Efficient Downstream Adaptation [51.67328507400985]
In noisy partial label learning (NPLL), each training sample is associated with a set of candidate labels annotated by multiple noisy annotators.<n>This paper focuses on learning from partial labels annotated by pre-trained vision-language models.<n>It proposes an innovative collaborative consistency regularization (Co-Reg) method.
arXiv Detail & Related papers (2025-06-03T12:48:54Z) - Adaptive Label Correction for Robust Medical Image Segmentation with Noisy Labels [18.868523370902338]
We propose a Mean Teacher-based Adaptive Label Correction framework for robust medical image segmentation with noisy labels.<n>It includes an adaptive label refinement mechanism that dynamically captures and weights differences across multiple disturbance versions to enhance the quality of noisy labels.<n>It also incorporates a sample-level uncertainty-based label selection algorithm to prioritize high-confidence samples for network updates.
arXiv Detail & Related papers (2025-03-15T18:03:01Z) - Set a Thief to Catch a Thief: Combating Label Noise through Noisy Meta Learning [6.68999525326685]
Learning from noisy labels (LNL) aims to train high-performance deep models using noisy datasets.<n>We propose a novel noisy meta label correction framework STCT, which counterintuitively uses noisy data to correct label noise.<n> STCT achieves 96.9% label correction and 95.2% classification performance on CIFAR-10 with 80% symmetric noise.
arXiv Detail & Related papers (2025-02-22T05:58:01Z) - Inaccurate Label Distribution Learning with Dependency Noise [52.08553913094809]
We introduce the Dependent Noise-based Inaccurate Label Distribution Learning (DN-ILDL) framework to tackle the challenges posed by noise in label distribution learning.
We show that DN-ILDL effectively addresses the ILDL problem and outperforms existing LDL methods.
arXiv Detail & Related papers (2024-05-26T07:58:07Z) - ERASE: Error-Resilient Representation Learning on Graphs for Label Noise
Tolerance [53.73316938815873]
We propose a method called ERASE (Error-Resilient representation learning on graphs for lAbel noiSe tolerancE) to learn representations with error tolerance.
ERASE combines prototype pseudo-labels with propagated denoised labels and updates representations with error resilience.
Our method can outperform multiple baselines with clear margins in broad noise levels and enjoy great scalability.
arXiv Detail & Related papers (2023-12-13T17:59:07Z) - Unleashing the Potential of Regularization Strategies in Learning with
Noisy Labels [65.92994348757743]
We demonstrate that a simple baseline using cross-entropy loss, combined with widely used regularization strategies can outperform state-of-the-art methods.
Our findings suggest that employing a combination of regularization strategies can be more effective than intricate algorithms in tackling the challenges of learning with noisy labels.
arXiv Detail & Related papers (2023-07-11T05:58:20Z) - Mitigating Label Noise through Data Ambiguation [9.51828574518325]
Large models with high expressive power are prone to memorizing incorrect labels, thereby harming generalization performance.
In this paper, we suggest to address the shortcomings of both methodologies by "ambiguating" the target information.
More precisely, we leverage the framework of so-called superset learning to construct set-valued targets based on a confidence threshold.
arXiv Detail & Related papers (2023-05-23T07:29:08Z) - Transductive CLIP with Class-Conditional Contrastive Learning [68.51078382124331]
We propose Transductive CLIP, a novel framework for learning a classification network with noisy labels from scratch.
A class-conditional contrastive learning mechanism is proposed to mitigate the reliance on pseudo labels.
ensemble labels is adopted as a pseudo label updating strategy to stabilize the training of deep neural networks with noisy labels.
arXiv Detail & Related papers (2022-06-13T14:04:57Z) - SELC: Self-Ensemble Label Correction Improves Learning with Noisy Labels [4.876988315151037]
Deep neural networks are prone to overfitting noisy labels, resulting in poor generalization performance.
We present a method self-ensemble label correction (SELC) to progressively correct noisy labels and refine the model.
SELC obtains more promising and stable results in the presence of class-conditional, instance-dependent, and real-world label noise.
arXiv Detail & Related papers (2022-05-02T18:42:47Z) - Learning with Out-of-Distribution Data for Audio Classification [60.48251022280506]
We show that detecting and relabelling certain OOD instances, rather than discarding them, can have a positive effect on learning.
The proposed method is shown to improve the performance of convolutional neural networks by a significant margin.
arXiv Detail & Related papers (2020-02-11T21:08:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.