Combating Noisy Labels in Long-Tailed Image Classification
- URL: http://arxiv.org/abs/2209.00273v1
- Date: Thu, 1 Sep 2022 07:31:03 GMT
- Title: Combating Noisy Labels in Long-Tailed Image Classification
- Authors: Chaowei Fang, Lechao Cheng, Huiyan Qi, and Dingwen Zhang
- Abstract summary: This paper makes an early effort to tackle the image classification task with both long-tailed distribution and label noise.
Existing noise-robust learning methods cannot work in this scenario as it is challenging to differentiate noisy samples from clean samples of tail classes.
We propose a new learning paradigm based on matching between inferences on weak and strong data augmentations to screen out noisy samples.
- Score: 33.40963778043824
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most existing methods that cope with noisy labels usually assume that the
class distributions are well balanced, which has insufficient capacity to deal
with the practical scenarios where training samples have imbalanced
distributions. To this end, this paper makes an early effort to tackle the
image classification task with both long-tailed distribution and label noise.
Existing noise-robust learning methods cannot work in this scenario as it is
challenging to differentiate noisy samples from clean samples of tail classes.
To deal with this problem, we propose a new learning paradigm based on matching
between inferences on weak and strong data augmentations to screen out noisy
samples and introduce a leave-noise-out regularization to eliminate the effect
of the recognized noisy samples. Furthermore, we incorporate a novel prediction
penalty based on online prior distribution to avoid bias towards head classes.
This mechanism has superiority in capturing the class fitting degree in
realtime compared to the existing long-tail classification methods. Exhaustive
experiments demonstrate that the proposed method outperforms state-of-the-art
algorithms that address the distribution imbalance problem in long-tailed
classification under noisy labels.
Related papers
- Learning from Noisy Labels for Long-tailed Data via Optimal Transport [2.8821062918162146]
We propose a novel approach to manage data characterized by both long-tailed distributions and noisy labels.
We employ optimal transport strategies to generate pseudo-labels for the noise set in a semi-supervised training manner.
arXiv Detail & Related papers (2024-08-07T14:15:18Z) - Active Label Refinement for Robust Training of Imbalanced Medical Image Classification Tasks in the Presence of High Label Noise [10.232537737211098]
We propose a two-phase approach that combines Learning with Noisy Labels (LNL) and active learning.
We demonstrate that our proposed technique is superior to its predecessors at handling class imbalance by not misidentifying clean samples from minority classes as mostly noisy samples.
arXiv Detail & Related papers (2024-07-08T14:16:05Z) - Extracting Clean and Balanced Subset for Noisy Long-tailed Classification [66.47809135771698]
We develop a novel pseudo labeling method using class prototypes from the perspective of distribution matching.
By setting a manually-specific probability measure, we can reduce the side-effects of noisy and long-tailed data simultaneously.
Our method can extract this class-balanced subset with clean labels, which brings effective performance gains for long-tailed classification with label noise.
arXiv Detail & Related papers (2024-04-10T07:34:37Z) - Learning with Imbalanced Noisy Data by Preventing Bias in Sample
Selection [82.43311784594384]
Real-world datasets contain not only noisy labels but also class imbalance.
We propose a simple yet effective method to address noisy labels in imbalanced datasets.
arXiv Detail & Related papers (2024-02-17T10:34:53Z) - Co-Learning Meets Stitch-Up for Noisy Multi-label Visual Recognition [70.00984078351927]
This paper focuses on reducing noise based on some inherent properties of multi-label classification and long-tailed learning under noisy cases.
We propose a Stitch-Up augmentation to synthesize a cleaner sample, which directly reduces multi-label noise.
A Heterogeneous Co-Learning framework is further designed to leverage the inconsistency between long-tailed and balanced distributions.
arXiv Detail & Related papers (2023-07-03T09:20:28Z) - Centrality and Consistency: Two-Stage Clean Samples Identification for
Learning with Instance-Dependent Noisy Labels [87.48541631675889]
We propose a two-stage clean samples identification method.
First, we employ a class-level feature clustering procedure for the early identification of clean samples.
Second, for the remaining clean samples that are close to the ground truth class boundary, we propose a novel consistency-based classification method.
arXiv Detail & Related papers (2022-07-29T04:54:57Z) - Robust Long-Tailed Learning under Label Noise [50.00837134041317]
This work investigates the label noise problem under long-tailed label distribution.
We propose a robust framework,algo, that realizes noise detection for long-tailed learning.
Our framework can naturally leverage semi-supervised learning algorithms to further improve the generalisation.
arXiv Detail & Related papers (2021-08-26T03:45:00Z) - Disentangling Sampling and Labeling Bias for Learning in Large-Output
Spaces [64.23172847182109]
We show that different negative sampling schemes implicitly trade-off performance on dominant versus rare labels.
We provide a unified means to explicitly tackle both sampling bias, arising from working with a subset of all labels, and labeling bias, which is inherent to the data due to label imbalance.
arXiv Detail & Related papers (2021-05-12T15:40:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.