MANTRA: A Machine Learning reference lightcurve dataset for astronomical
transient event recognition
- URL: http://arxiv.org/abs/2006.13163v2
- Date: Tue, 30 Jun 2020 17:45:58 GMT
- Title: MANTRA: A Machine Learning reference lightcurve dataset for astronomical
transient event recognition
- Authors: Mauricio Neira, Catalina G\'omez, John F. Su\'arez-P\'erez, Diego A.
G\'omez, Juan Pablo Reyes, Marcela Hern\'andez Hoyos, Pablo Arbel\'aez, Jaime
E. Forero-Romero
- Abstract summary: We provide public access to a dataset of 4869 transient and 71207 non-transient object lightcurves built from the Catalina Real Time Transient Survey.
Some of the classes included in the dataset are: supernovae, cataclysmic variables, active galactic nuclei, high proper motion stars, blazars and flares.
We assess quantitative performance in two classification tasks: binary (transient/non-transient) and eight-class classification.
- Score: 2.208166456405677
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce MANTRA, an annotated dataset of 4869 transient and 71207
non-transient object lightcurves built from the Catalina Real Time Transient
Survey. We provide public access to this dataset as a plain text file to
facilitate standardized quantitative comparison of astronomical transient event
recognition algorithms. Some of the classes included in the dataset are:
supernovae, cataclysmic variables, active galactic nuclei, high proper motion
stars, blazars and flares. As an example of the tasks that can be performed on
the dataset we experiment with multiple data pre-processing methods, feature
selection techniques and popular machine learning algorithms (Support Vector
Machines, Random Forests and Neural Networks). We assess quantitative
performance in two classification tasks: binary (transient/non-transient) and
eight-class classification. The best performing algorithm in both tasks is the
Random Forest Classifier. It achieves an F1-score of 96.25% in the binary
classification and 52.79% in the eight-class classification. For the
eight-class classification, non-transients ( 96.83% ) is the class with the
highest F1-score, while the lowest corresponds to high-proper-motion stars (
16.79% ); for supernovae it achieves a value of 54.57% , close to the average
across classes. The next release of MANTRA includes images and benchmarks with
deep learning models.
Related papers
- ORACLE: A Real-Time, Hierarchical, Deep-Learning Photometric Classifier for the LSST [0.3276793654637396]
We present ORACLE, the first hierarchical deep-learning model for real-time, context-aware classification of transient and variable astrophysical phenomena.
ORACLE is a recurrent neural network with Gated Recurrent Units (GRUs)
Training on $sim$0.5M events from the Extended LSST Astronomical Time-Series Classification Challenge, we achieve a top-level (Transient vs Variable) macro-averaged precision of 0.96 using only 1 day of photometric observations.
arXiv Detail & Related papers (2025-01-02T19:00:05Z) - A Novel Adaptive Fine-Tuning Algorithm for Multimodal Models: Self-Optimizing Classification and Selection of High-Quality Datasets in Remote Sensing [46.603157010223505]
We propose an adaptive fine-tuning algorithm for multimodal large models.
We train the model on two 3090 GPU using one-third of the GeoChat multimodal remote sensing dataset.
The model achieved scores of 89.86 and 77.19 on the UCMerced and AID evaluation datasets.
arXiv Detail & Related papers (2024-09-20T09:19:46Z) - Finding active galactic nuclei through Fink [0.0]
We present the Active Galactic Nuclei (AGN) classifier as currently implemented within the Fink broker.
Features were built upon summary statistics of available photometric points, as well as color estimation enabled by symbolic regression.
Using this method to classify real alerts from the Zwicky Transient Facility (ZTF), we achieved 98.0% accuracy, 93.8% precision and 88.5% recall.
arXiv Detail & Related papers (2022-11-20T14:24:15Z) - Semi-Supervised Domain Adaptation for Cross-Survey Galaxy Morphology
Classification and Anomaly Detection [57.85347204640585]
We develop a Universal Domain Adaptation method DeepAstroUDA.
It can be applied to datasets with different types of class overlap.
For the first time, we demonstrate the successful use of domain adaptation on two very different observational datasets.
arXiv Detail & Related papers (2022-11-01T18:07:21Z) - SNGuess: A method for the selection of young extragalactic transients [0.0]
This paper presents SNGuess, a model designed to find young extragalactic nearby transients with high purity.
SNGuess works with a set of features that can be efficiently calculated from astronomical alert data.
The core model of SNGuess consists of an ensemble of decision trees, which are trained via gradient boosting.
arXiv Detail & Related papers (2022-08-13T00:11:46Z) - Fink: early supernovae Ia classification using active learning [0.0]
We demonstrate the feasibility of implementation of such strategies in the current Zwicky Transient Facility (ZTF) public alert data stream.
Our results confirm the effectiveness of active learning strategies for guiding the construction of optimal training samples for astronomical classifiers.
arXiv Detail & Related papers (2021-11-22T11:38:58Z) - SCARF: Self-Supervised Contrastive Learning using Random Feature
Corruption [72.35532598131176]
We propose SCARF, a technique for contrastive learning, where views are formed by corrupting a random subset of features.
We show that SCARF complements existing strategies and outperforms alternatives like autoencoders.
arXiv Detail & Related papers (2021-06-29T08:08:33Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - StarcNet: Machine Learning for Star Cluster Identification [21.804465402347027]
StarcNet is a convolutional neural network (CNN) which achieves an accuracy of 68.6% (4 classes)/86.0% (2 classes: cluster/non-cluster) for star cluster classification in the images of the LEGUS galaxies, nearly matching human expert performance.
We test the performance of StarcNet by applying pre-trained CNN model to galaxies not included in the training set, finding accuracies similar to the reference one.
In luminosity, color, and physical characteristics of star clusters are similar for the human and ML classified samples.
arXiv Detail & Related papers (2020-12-16T23:58:01Z) - Device-Robust Acoustic Scene Classification Based on Two-Stage
Categorization and Data Augmentation [63.98724740606457]
We present a joint effort of four groups, namely GT, USTC, Tencent, and UKE, to tackle Task 1 - Acoustic Scene Classification (ASC) in the DCASE 2020 Challenge.
Task 1a focuses on ASC of audio signals recorded with multiple (real and simulated) devices into ten different fine-grained classes.
Task 1b concerns with classification of data into three higher-level classes using low-complexity solutions.
arXiv Detail & Related papers (2020-07-16T15:07:14Z) - SCAN: Learning to Classify Images without Labels [73.69513783788622]
We advocate a two-step approach where feature learning and clustering are decoupled.
A self-supervised task from representation learning is employed to obtain semantically meaningful features.
We obtain promising results on ImageNet, and outperform several semi-supervised learning methods in the low-data regime.
arXiv Detail & Related papers (2020-05-25T18:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.