Suppressing Mislabeled Data via Grouping and Self-Attention
- URL: http://arxiv.org/abs/2010.15603v1
- Date: Thu, 29 Oct 2020 13:54:16 GMT
- Title: Suppressing Mislabeled Data via Grouping and Self-Attention
- Authors: Xiaojiang Peng, Kai Wang, Zhaoyang Zeng, Qing Li, Jianfei Yang and Yu
Qiao
- Abstract summary: Deep networks achieve excellent results on large-scale clean data but degrade significantly when learning from noisy labels.
This paper proposes a conceptually simple yet efficient training block, termed as Attentive Feature Mixup (AFM)
It allows paying more attention to clean samples and less to mislabeled ones via sample interactions in small groups.
- Score: 60.14212694011875
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep networks achieve excellent results on large-scale clean data but degrade
significantly when learning from noisy labels. To suppressing the impact of
mislabeled data, this paper proposes a conceptually simple yet efficient
training block, termed as Attentive Feature Mixup (AFM), which allows paying
more attention to clean samples and less to mislabeled ones via sample
interactions in small groups. Specifically, this plug-and-play AFM first
leverages a \textit{group-to-attend} module to construct groups and assign
attention weights for group-wise samples, and then uses a \textit{mixup} module
with the attention weights to interpolate massive noisy-suppressed samples. The
AFM has several appealing benefits for noise-robust deep learning. (i) It does
not rely on any assumptions and extra clean subset. (ii) With massive
interpolations, the ratio of useless samples is reduced dramatically compared
to the original noisy ratio. (iii) \pxj{It jointly optimizes the interpolation
weights with classifiers, suppressing the influence of mislabeled data via low
attention weights. (iv) It partially inherits the vicinal risk minimization of
mixup to alleviate over-fitting while improves it by sampling fewer
feature-target vectors around mislabeled data from the mixup vicinal
distribution.} Extensive experiments demonstrate that AFM yields
state-of-the-art results on two challenging real-world noisy datasets: Food101N
and Clothing1M. The code will be available at
https://github.com/kaiwang960112/AFM.
Related papers
- Mitigating Noisy Supervision Using Synthetic Samples with Soft Labels [13.314778587751588]
Noisy labels are ubiquitous in real-world datasets, especially in the large-scale ones derived from crowdsourcing and web searching.
It is challenging to train deep neural networks with noisy datasets since the networks are prone to overfitting the noisy labels during training.
We propose a framework that trains the model with new synthetic samples to mitigate the impact of noisy labels.
arXiv Detail & Related papers (2024-06-22T04:49:39Z) - Extracting Clean and Balanced Subset for Noisy Long-tailed Classification [66.47809135771698]
We develop a novel pseudo labeling method using class prototypes from the perspective of distribution matching.
By setting a manually-specific probability measure, we can reduce the side-effects of noisy and long-tailed data simultaneously.
Our method can extract this class-balanced subset with clean labels, which brings effective performance gains for long-tailed classification with label noise.
arXiv Detail & Related papers (2024-04-10T07:34:37Z) - Manifold DivideMix: A Semi-Supervised Contrastive Learning Framework for
Severe Label Noise [4.90148689564172]
Real-world datasets contain noisy label samples that have no semantic relevance to any class in the dataset.
Most state-of-the-art methods leverage ID labeled noisy samples as unlabeled data for semi-supervised learning.
We propose incorporating the information from all the training data by leveraging the benefits of self-supervised training.
arXiv Detail & Related papers (2023-08-13T23:33:33Z) - MAPS: A Noise-Robust Progressive Learning Approach for Source-Free
Domain Adaptive Keypoint Detection [76.97324120775475]
Cross-domain keypoint detection methods always require accessing the source data during adaptation.
This paper considers source-free domain adaptive keypoint detection, where only the well-trained source model is provided to the target domain.
arXiv Detail & Related papers (2023-02-09T12:06:08Z) - Learning from Noisy Labels with Coarse-to-Fine Sample Credibility
Modeling [22.62790706276081]
Training deep neural network (DNN) with noisy labels is practically challenging.
Previous efforts tend to handle part or full data in a unified denoising flow.
We propose a coarse-to-fine robust learning method called CREMA to handle noisy data in a divide-and-conquer manner.
arXiv Detail & Related papers (2022-08-23T02:06:38Z) - Neighborhood Collective Estimation for Noisy Label Identification and
Correction [92.20697827784426]
Learning with noisy labels (LNL) aims at designing strategies to improve model performance and generalization by mitigating the effects of model overfitting to noisy labels.
Recent advances employ the predicted label distributions of individual samples to perform noise verification and noisy label correction, easily giving rise to confirmation bias.
We propose Neighborhood Collective Estimation, in which the predictive reliability of a candidate sample is re-estimated by contrasting it against its feature-space nearest neighbors.
arXiv Detail & Related papers (2022-08-05T14:47:22Z) - Boosting Factorization Machines via Saliency-Guided Mixup [125.15872106335692]
We present MixFM, inspired by Mixup, to generate auxiliary training data to boost Factorization machines (FMs)
We also put forward a novel Factorization Machine powered by Saliency-guided Mixup (denoted as SMFM)
arXiv Detail & Related papers (2022-06-17T09:49:00Z) - Harnessing Hard Mixed Samples with Decoupled Regularizer [69.98746081734441]
Mixup is an efficient data augmentation approach that improves the generalization of neural networks by smoothing the decision boundary with mixed data.
In this paper, we propose an efficient mixup objective function with a decoupled regularizer named Decoupled Mixup (DM)
DM can adaptively utilize hard mixed samples to mine discriminative features without losing the original smoothness of mixup.
arXiv Detail & Related papers (2022-03-21T07:12:18Z) - Noisy Feature Mixup [42.056684988818766]
We introduce Noisy Feature Mixup (NFM), an inexpensive yet effective method for data augmentation.
NFM includes mixup and manifold mixup as special cases, but it has additional advantages, including better smoothing of decision boundaries.
We show that residual networks and vision transformers trained with NFM have favorable trade-offs between predictive accuracy on clean data and robustness with respect to various types of data.
arXiv Detail & Related papers (2021-10-05T17:13:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.