Fink: early supernovae Ia classification using active learning
- URL: http://arxiv.org/abs/2111.11438v1
- Date: Mon, 22 Nov 2021 11:38:58 GMT
- Title: Fink: early supernovae Ia classification using active learning
- Authors: Marco Leoni, Emille E. O. Ishida, Julien Peloton and Anais M\"oller
- Abstract summary: We demonstrate the feasibility of implementation of such strategies in the current Zwicky Transient Facility (ZTF) public alert data stream.
Our results confirm the effectiveness of active learning strategies for guiding the construction of optimal training samples for astronomical classifiers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We describe how the Fink broker early supernova Ia classifier optimizes its
ML classifications by employing an active learning (AL) strategy. We
demonstrate the feasibility of implementation of such strategies in the current
Zwicky Transient Facility (ZTF) public alert data stream. We compare the
performance of two AL strategies: uncertainty sampling and random sampling. Our
pipeline consists of 3 stages: feature extraction, classification and learning
strategy. Starting from an initial sample of 10 alerts (5 SN Ia and 5 non-Ia),
we let the algorithm identify which alert should be added to the training
sample. The system is allowed to evolve through 300 iterations. Our data set
consists of 23 840 alerts from the ZTF with confirmed classification via
cross-match with SIMBAD database and the Transient name server (TNS), 1 600 of
which were SNe Ia (1 021 unique objects). The data configuration, after the
learning cycle was completed, consists of 310 alerts for training and 23 530
for testing. Averaging over 100 realizations, the classifier achieved 89%
purity and 54% efficiency. From 01/November/2020 to 31/October/2021 Fink has
applied its early supernova Ia module to the ZTF stream and communicated
promising SN Ia candidates to the TNS. From the 535 spectroscopically
classified Fink candidates, 459 (86%) were proven to be SNe Ia. Our results
confirm the effectiveness of active learning strategies for guiding the
construction of optimal training samples for astronomical classifiers. It
demonstrates in real data that the performance of learning algorithms can be
highly improved without the need of extra computational resources or
overwhelmingly large training samples. This is, to our knowledge, the first
application of AL to real alerts data.
Related papers
- Class Balance Matters to Active Class-Incremental Learning [61.11786214164405]
We aim to start from a pool of large-scale unlabeled data and then annotate the most informative samples for incremental learning.
We propose Class-Balanced Selection (CBS) strategy to achieve both class balance and informativeness in chosen samples.
Our CBS can be plugged and played into those CIL methods which are based on pretrained models with prompts tunning technique.
arXiv Detail & Related papers (2024-12-09T16:37:27Z) - DST-Det: Simple Dynamic Self-Training for Open-Vocabulary Object Detection [72.25697820290502]
This work introduces a straightforward and efficient strategy to identify potential novel classes through zero-shot classification.
We refer to this approach as the self-training strategy, which enhances recall and accuracy for novel classes without requiring extra annotations, datasets, and re-training.
Empirical evaluations on three datasets, including LVIS, V3Det, and COCO, demonstrate significant improvements over the baseline performance.
arXiv Detail & Related papers (2023-10-02T17:52:24Z) - Convolutional Neural Networks for the classification of glitches in
gravitational-wave data streams [52.77024349608834]
We classify transient noise signals (i.e.glitches) and gravitational waves in data from the Advanced LIGO detectors.
We use models with a supervised learning approach, both trained from scratch using the Gravity Spy dataset.
We also explore a self-supervised approach, pre-training models with automatically generated pseudo-labels.
arXiv Detail & Related papers (2023-03-24T11:12:37Z) - Segment-level Metric Learning for Few-shot Bioacoustic Event Detection [56.59107110017436]
We propose a segment-level few-shot learning framework that utilizes both the positive and negative events during model optimization.
Our system achieves an F-measure of 62.73 on the DCASE 2022 challenge task 5 (DCASE2022-T5) validation set, outperforming the performance of the baseline prototypical network 34.02 by a large margin.
arXiv Detail & Related papers (2022-07-15T22:41:30Z) - CNLL: A Semi-supervised Approach For Continual Noisy Label Learning [12.341250124228859]
We propose a simple purification technique to effectively cleanse the online data stream that is both cost-effective and more accurate.
After purification, we perform fine-tuning in a semi-supervised fashion that ensures the participation of all available samples.
We achieve a 24.8% performance gain for CIFAR10 with 20% noise over previous SOTA methods.
arXiv Detail & Related papers (2022-04-21T05:01:10Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - Few-Shot Incremental Learning with Continually Evolved Classifiers [46.278573301326276]
Few-shot class-incremental learning (FSCIL) aims to design machine learning algorithms that can continually learn new concepts from a few data points.
The difficulty lies in that limited data from new classes not only lead to significant overfitting issues but also exacerbate the notorious catastrophic forgetting problems.
We propose a Continually Evolved CIF ( CEC) that employs a graph model to propagate context information between classifiers for adaptation.
arXiv Detail & Related papers (2021-04-07T10:54:51Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Alert Classification for the ALeRCE Broker System: The Real-time Stamp
Classifier [0.0]
We present a real-time stamp of astronomical events for the ALeRCE broker.
Our work represents an important milestone toward rapid alert classifications with the next generation of large etendue telescopes.
arXiv Detail & Related papers (2020-08-07T18:00:01Z) - MANTRA: A Machine Learning reference lightcurve dataset for astronomical
transient event recognition [2.208166456405677]
We provide public access to a dataset of 4869 transient and 71207 non-transient object lightcurves built from the Catalina Real Time Transient Survey.
Some of the classes included in the dataset are: supernovae, cataclysmic variables, active galactic nuclei, high proper motion stars, blazars and flares.
We assess quantitative performance in two classification tasks: binary (transient/non-transient) and eight-class classification.
arXiv Detail & Related papers (2020-06-23T17:06:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.