Supernova Light Curves Approximation based on Neural Network Models
- URL: http://arxiv.org/abs/2206.13306v1
- Date: Mon, 27 Jun 2022 13:46:51 GMT
- Title: Supernova Light Curves Approximation based on Neural Network Models
- Authors: Mariia Demianenko, Ekaterina Samorodova, Mikhail Sysak, Aleksandr
Shiriaev, Konstantin Malanchev, Denis Derkach, Mikhail Hushchyn
- Abstract summary: Photometric data-driven classification of supernovae becomes a challenge due to the appearance of real-time processing of big data in astronomy.
Recent studies have demonstrated the superior quality of solutions based on various machine learning models.
We study the application of multilayer perceptron (MLP), bayesian neural network (BNN), and normalizing flows (NF) to approximate observations for a single light curve.
- Score: 53.180678723280145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Photometric data-driven classification of supernovae becomes a challenge due
to the appearance of real-time processing of big data in astronomy. Recent
studies have demonstrated the superior quality of solutions based on various
machine learning models. These models learn to classify supernova types using
their light curves as inputs. Preprocessing these curves is a crucial step that
significantly affects the final quality. In this talk, we study the application
of multilayer perceptron (MLP), bayesian neural network (BNN), and normalizing
flows (NF) to approximate observations for a single light curve. We use these
approximations as inputs for supernovae classification models and demonstrate
that the proposed methods outperform the state-of-the-art based on Gaussian
processes applying to the Zwicky Transient Facility Bright Transient Survey
light curves. MLP demonstrates similar quality as Gaussian processes and speed
increase. Normalizing Flows exceeds Gaussian processes in terms of
approximation quality as well.
Related papers
- Deep Learning and LLM-based Methods Applied to Stellar Lightcurve Classification [7.592813175419603]
We present a comprehensive evaluation of deep-learning and large language model (LLM) based models for the automatic classification of variable star light curves.
Special emphasis is placed on Cepheids, RR Lyrae, and eclipsing binaries, examining the influence of observational cadence and phase distribution on classification precision.
We unveil StarWhisper LightCurve (LC), an innovative Series comprising three LLM-based models: LLM, multimodal large language model (MLLM), and Large Audio Language Model (LALM)
arXiv Detail & Related papers (2024-04-16T17:35:25Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - GAFlow: Incorporating Gaussian Attention into Optical Flow [62.646389181507764]
We push Gaussian Attention (GA) into the optical flow models to accentuate local properties during representation learning.
We introduce a novel Gaussian-Constrained Layer (GCL) which can be easily plugged into existing Transformer blocks.
For reliable motion analysis, we provide a new Gaussian-Guided Attention Module (GGAM)
arXiv Detail & Related papers (2023-09-28T07:46:01Z) - Neural Inference of Gaussian Processes for Time Series Data of Quasars [72.79083473275742]
We introduce a new model that enables it to describe quasar spectra completely.
We also introduce a new method of inference of Gaussian process parameters, which we call $textitNeural Inference$.
The combination of both the CDRW model and Neural Inference significantly outperforms the baseline DRW and MLE.
arXiv Detail & Related papers (2022-11-17T13:01:26Z) - Understanding of the properties of neural network approaches for
transient light curve approximations [37.91290708320157]
This paper presents a search for the best-performing methods to approximate the observed light curves over time and wavelength.
Test datasets include simulated PLAsTiCC and real Zwicky Transient Facility Bright Transient Survey light curves of transients.
arXiv Detail & Related papers (2022-09-15T18:00:08Z) - Light curve completion and forecasting using fast and scalable Gaussian
processes (MuyGPs) [0.0]
Ground-based observations from commercial off the shelf (COTS) cameras remain inexpensive compared to higher precision instruments.
limited sensor availability combined with noisier observations can produce gappy time-series data.
Deep Neural Networks (DNNs) have become the tool of choice due to their empirical success at learning complex nonlinear embeddings.
arXiv Detail & Related papers (2022-08-31T01:52:00Z) - A Convolutional Neural Network Approach to Supernova Time-Series
Classification [0.0]
We present a convolutional neural network method for fast supernova time-series classification.
We are able to differentiate between 6 SN types with 60% accuracy given only two nights of data and 98% accuracy retrospectively.
arXiv Detail & Related papers (2022-07-19T17:55:22Z) - Inferring Structural Parameters of Low-Surface-Brightness-Galaxies with
Uncertainty Quantification using Bayesian Neural Networks [70.80563014913676]
We show that a Bayesian Neural Network (BNN) can be used for the inference, with uncertainty, of such parameters from simulated low-surface-brightness galaxy images.
Compared to traditional profile-fitting methods, we show that the uncertainties obtained using BNNs are comparable in magnitude, well-calibrated, and the point estimates of the parameters are closer to the true values.
arXiv Detail & Related papers (2022-07-07T17:55:26Z) - Local Random Feature Approximations of the Gaussian Kernel [14.230653042112834]
We focus on the popular Gaussian kernel and on techniques to linearize kernel-based models by means of random feature approximations.
We show that such approaches yield poor results when modelling high-frequency data, and we propose a novel localization scheme that improves kernel approximations and downstream performance significantly.
arXiv Detail & Related papers (2022-04-12T09:52:36Z) - Kernel and Rich Regimes in Overparametrized Models [69.40899443842443]
We show that gradient descent on overparametrized multilayer networks can induce rich implicit biases that are not RKHS norms.
We also demonstrate this transition empirically for more complex matrix factorization models and multilayer non-linear networks.
arXiv Detail & Related papers (2020-02-20T15:43:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.