Noisy Labels for Weakly Supervised Gamma Hadron Classification
- URL: http://arxiv.org/abs/2108.13396v1
- Date: Mon, 30 Aug 2021 17:38:20 GMT
- Title: Noisy Labels for Weakly Supervised Gamma Hadron Classification
- Authors: Lukas Pfahler, Mirko Bunse, Katharina Morik
- Abstract summary: We propose to solve gamma hadron classification with a noisy label approach that only uses unlabeled data recorded by the real telescope.
We show that models which are based on the significance of detection deliver state-of-the-art results, despite being exclusively trained with noisy labels.
- Score: 9.249443355045967
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gamma hadron classification, a central machine learning task in gamma ray
astronomy, is conventionally tackled with supervised learning. However, the
supervised approach requires annotated training data to be produced in
sophisticated and costly simulations. We propose to instead solve gamma hadron
classification with a noisy label approach that only uses unlabeled data
recorded by the real telescope. To this end, we employ the significance of
detection as a learning criterion which addresses this form of weak
supervision. We show that models which are based on the significance of
detection deliver state-of-the-art results, despite being exclusively trained
with noisy labels; put differently, our models do not require the costly
simulated ground-truth labels that astronomers otherwise employ for classifier
training. Our weakly supervised models exhibit competitive performances also on
imbalanced data sets that stem from a variety of other application domains. In
contrast to existing work on class-conditional label noise, we assume that only
one of the class-wise noise rates is known.
Related papers
- Learning Discriminative Dynamics with Label Corruption for Noisy Label Detection [25.55455239006278]
We propose DynaCor framework that distinguishes incorrectly labeled instances from correctly labeled ones based on the dynamics of the training signals.
Our comprehensive experiments show that DynaCor outperforms the state-of-the-art competitors and shows strong robustness to various noise types and noise rates.
arXiv Detail & Related papers (2024-05-30T10:06:06Z) - Virtual Category Learning: A Semi-Supervised Learning Method for Dense
Prediction with Extremely Limited Labels [63.16824565919966]
This paper proposes to use confusing samples proactively without label correction.
A Virtual Category (VC) is assigned to each confusing sample in such a way that it can safely contribute to the model optimisation.
Our intriguing findings highlight the usage of VC learning in dense vision tasks.
arXiv Detail & Related papers (2023-12-02T16:23:52Z) - Soft Curriculum for Learning Conditional GANs with Noisy-Labeled and
Uncurated Unlabeled Data [70.25049762295193]
We introduce a novel conditional image generation framework that accepts noisy-labeled and uncurated data during training.
We propose soft curriculum learning, which assigns instance-wise weights for adversarial training while assigning new labels for unlabeled data.
Our experiments show that our approach outperforms existing semi-supervised and label-noise robust methods in terms of both quantitative and qualitative performance.
arXiv Detail & Related papers (2023-07-17T08:31:59Z) - Learning with Noisy Labels through Learnable Weighting and Centroid Similarity [5.187216033152917]
noisy labels are prevalent in domains such as medical diagnosis and autonomous driving.
We introduce a novel method for training machine learning models in the presence of noisy labels.
Our results show that our method consistently outperforms the existing state-of-the-art techniques.
arXiv Detail & Related papers (2023-03-16T16:43:24Z) - Semi-supervised Object Detection via Virtual Category Learning [68.26956850996976]
This paper proposes to use confusing samples proactively without label correction.
Specifically, a virtual category (VC) is assigned to each confusing sample.
It is attributed to specifying the embedding distance between the training sample and the virtual category.
arXiv Detail & Related papers (2022-07-07T16:59:53Z) - S3: Supervised Self-supervised Learning under Label Noise [53.02249460567745]
In this paper we address the problem of classification in the presence of label noise.
In the heart of our method is a sample selection mechanism that relies on the consistency between the annotated label of a sample and the distribution of the labels in its neighborhood in the feature space.
Our method significantly surpasses previous methods on both CIFARCIFAR100 with artificial noise and real-world noisy datasets such as WebVision and ANIMAL-10N.
arXiv Detail & Related papers (2021-11-22T15:49:20Z) - Unsupervised Learning Architecture for Classifying the Transient Noise
of Interferometric Gravitational-wave Detectors [2.8555963243398073]
transient noise with non-stationary and non-Gaussian features occurs at a high rate.
Classification of transient noise can offer clues for exploring its origin and improving the performance of the detector.
In this study, we propose an unsupervised learning architecture for the classification of transient noise.
arXiv Detail & Related papers (2021-11-19T05:37:06Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - Ensemble Learning with Manifold-Based Data Splitting for Noisy Label
Correction [20.401661156102897]
noisy labels in training data can significantly degrade a model's generalization performance.
We propose an ensemble learning method to correct noisy labels by exploiting the local structures of feature manifold.
Our experiments on real-world noisy label datasets demonstrate the superiority of the proposed method over existing state-of-the-arts.
arXiv Detail & Related papers (2021-03-13T07:24:58Z) - Robustness of Accuracy Metric and its Inspirations in Learning with
Noisy Labels [51.66448070984615]
We show that maximizing training accuracy on sufficiently many noisy samples yields an approximately optimal classifier.
For validation, we prove that a noisy validation set is reliable, addressing the critical demand of model selection.
We show characterizations of models trained with noisy labels, motivated by our theoretical results, and verify the utility of a noisy validation set.
arXiv Detail & Related papers (2020-12-08T03:37:47Z) - A Survey on Deep Learning with Noisy Labels: How to train your model
when you cannot trust on the annotations? [21.562089974755125]
Several approaches have been proposed to improve the training of deep learning models in the presence of noisy labels.
This paper presents a survey on the main techniques in literature, in which we classify the algorithm in the following groups: robust losses, sample weighting, sample selection, meta-learning, and combined approaches.
arXiv Detail & Related papers (2020-12-05T15:45:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.