DOST -- Domain Obedient Self-supervised Training for Multi Label
Classification with Noisy Labels
- URL: http://arxiv.org/abs/2308.05101v1
- Date: Wed, 9 Aug 2023 17:53:36 GMT
- Title: DOST -- Domain Obedient Self-supervised Training for Multi Label
Classification with Noisy Labels
- Authors: Soumadeep Saha, Utpal Garain, Arijit Ukil, Arpan Pal, Sundeep
Khandelwal
- Abstract summary: This paper studies the effect of label noise on domain rule violation incidents in the multi-label classification task.
We propose the Domain Obedient Self-supervised Training (DOST) paradigm which makes deep learning models more aligned to domain rules.
- Score: 27.696103256353254
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The enormous demand for annotated data brought forth by deep learning
techniques has been accompanied by the problem of annotation noise. Although
this issue has been widely discussed in machine learning literature, it has
been relatively unexplored in the context of "multi-label classification" (MLC)
tasks which feature more complicated kinds of noise. Additionally, when the
domain in question has certain logical constraints, noisy annotations often
exacerbate their violations, making such a system unacceptable to an expert.
This paper studies the effect of label noise on domain rule violation incidents
in the MLC task, and incorporates domain rules into our learning algorithm to
mitigate the effect of noise. We propose the Domain Obedient Self-supervised
Training (DOST) paradigm which not only makes deep learning models more aligned
to domain rules, but also improves learning performance in key metrics and
minimizes the effect of annotation noise. This novel approach uses domain
guidance to detect offending annotations and deter rule-violating predictions
in a self-supervised manner, thus making it more "data efficient" and domain
compliant. Empirical studies, performed over two large scale multi-label
classification datasets, demonstrate that our method results in improvement
across the board, and often entirely counteracts the effect of noise.
Related papers
- Correcting Noisy Multilabel Predictions: Modeling Label Noise through Latent Space Shifts [4.795811957412855]
Noise in data appears to be inevitable in most real-world machine learning applications.
We investigate the less explored area of noisy label learning for multilabel classifications.
Our model posits that label noise arises from a shift in the latent variable, providing a more robust and beneficial means for noisy learning.
arXiv Detail & Related papers (2025-02-20T05:41:52Z) - Mitigating Instance-Dependent Label Noise: Integrating Self-Supervised Pretraining with Pseudo-Label Refinement [3.272177633069322]
Real-world datasets often contain noisy labels due to human error, ambiguity, or resource constraints during the annotation process.
We propose a novel framework that combines self-supervised learning using SimCLR with iterative pseudo-label refinement.
Our approach significantly outperforms several state-of-the-art methods, particularly under high noise conditions.
arXiv Detail & Related papers (2024-12-06T09:56:49Z) - SoftPatch: Unsupervised Anomaly Detection with Noisy Data [67.38948127630644]
This paper considers label-level noise in image sensory anomaly detection for the first time.
We propose a memory-based unsupervised AD method, SoftPatch, which efficiently denoises the data at the patch level.
Compared with existing methods, SoftPatch maintains a strong modeling ability of normal data and alleviates the overconfidence problem in coreset.
arXiv Detail & Related papers (2024-03-21T08:49:34Z) - Robust Tiny Object Detection in Aerial Images amidst Label Noise [50.257696872021164]
This study addresses the issue of tiny object detection under noisy label supervision.
We propose a DeNoising Tiny Object Detector (DN-TOD), which incorporates a Class-aware Label Correction scheme.
Our method can be seamlessly integrated into both one-stage and two-stage object detection pipelines.
arXiv Detail & Related papers (2024-01-16T02:14:33Z) - Co-Learning Meets Stitch-Up for Noisy Multi-label Visual Recognition [70.00984078351927]
This paper focuses on reducing noise based on some inherent properties of multi-label classification and long-tailed learning under noisy cases.
We propose a Stitch-Up augmentation to synthesize a cleaner sample, which directly reduces multi-label noise.
A Heterogeneous Co-Learning framework is further designed to leverage the inconsistency between long-tailed and balanced distributions.
arXiv Detail & Related papers (2023-07-03T09:20:28Z) - Rethinking the Value of Labels for Instance-Dependent Label Noise
Learning [43.481591776038144]
noisy labels in real-world applications often depend on both the true label and the features.
In this work, we tackle instance-dependent label noise with a novel deep generative model that avoids explicitly modeling the noise transition matrix.
Our algorithm leverages casual representation learning and simultaneously identifies the high-level content and style latent factors from the data.
arXiv Detail & Related papers (2023-05-10T15:29:07Z) - Noisy Annotation Refinement for Object Detection [47.066070566714984]
We propose a new problem setting of training object detectors on datasets with entangled noises of annotations of class labels and bounding boxes.
Our proposed method efficiently decouples the entangled noises, corrects the noisy annotations, and subsequently trains the detector using the corrected annotations.
arXiv Detail & Related papers (2021-10-20T09:39:50Z) - Open-set Label Noise Can Improve Robustness Against Inherent Label Noise [27.885927200376386]
We show that open-set noisy labels can be non-toxic and even benefit the robustness against inherent noisy labels.
We propose a simple yet effective regularization by introducing Open-set samples with Dynamic Noisy Labels (ODNL) into training.
arXiv Detail & Related papers (2021-06-21T07:15:50Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - Attention-Aware Noisy Label Learning for Image Classification [97.26664962498887]
Deep convolutional neural networks (CNNs) learned on large-scale labeled samples have achieved remarkable progress in computer vision.
The cheapest way to obtain a large body of labeled visual data is to crawl from websites with user-supplied labels, such as Flickr.
This paper proposes the attention-aware noisy label learning approach to improve the discriminative capability of the network trained on datasets with potential label noise.
arXiv Detail & Related papers (2020-09-30T15:45:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.