A Probabilistic Autoencoder for Type Ia Supernovae Spectral Time Series
- URL: http://arxiv.org/abs/2207.07645v1
- Date: Fri, 15 Jul 2022 17:58:27 GMT
- Title: A Probabilistic Autoencoder for Type Ia Supernovae Spectral Time Series
- Authors: George Stein, Uros Seljak, Vanessa Bohm, G. Aldering, P. Antilogus, C.
Aragon, S. Bailey, C. Baltay, S. Bongard, K. Boone, C. Buton, Y. Copin, S.
Dixon, D. Fouchez, E. Gangler, R. Gupta, B. Hayden, W. Hillebrandt, M.
Karmen, A. G. Kim, M. Kowalski, D. Kusters, P. F. Leget, F. Mondon, J.
Nordin, R. Pain, E. Pecontal, R. Pereira, S. Perlmutter, K. A. Ponder, D.
Rabinowitz, M. Rigault, D. Rubin, K. Runge, C. Saunders, G. Smadja, N.
Suzuki, C. Tao, R. C. Thomas, M. Vincenzi
- Abstract summary: We construct a probabilistic autoencoder to learn the intrinsic diversity of type Ia supernovae (SNe Ia) from a sparse set of spectral time series.
We show that the PAE learns a low-dimensional latent space that captures the nonlinear range of features that exists within the population.
We then use our PAE in a number of downstream tasks on SNe Ia for increasingly precise cosmological analyses.
- Score: 1.3316902142331577
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We construct a physically-parameterized probabilistic autoencoder (PAE) to
learn the intrinsic diversity of type Ia supernovae (SNe Ia) from a sparse set
of spectral time series. The PAE is a two-stage generative model, composed of
an Auto-Encoder (AE) which is interpreted probabilistically after training
using a Normalizing Flow (NF). We demonstrate that the PAE learns a
low-dimensional latent space that captures the nonlinear range of features that
exists within the population, and can accurately model the spectral evolution
of SNe Ia across the full range of wavelength and observation times directly
from the data. By introducing a correlation penalty term and multi-stage
training setup alongside our physically-parameterized network we show that
intrinsic and extrinsic modes of variability can be separated during training,
removing the need for the additional models to perform magnitude
standardization. We then use our PAE in a number of downstream tasks on SNe Ia
for increasingly precise cosmological analyses, including automatic detection
of SN outliers, the generation of samples consistent with the data
distribution, and solving the inverse problem in the presence of noisy and
incomplete data to constrain cosmological distance measurements. We find that
the optimal number of intrinsic model parameters appears to be three, in line
with previous studies, and show that we can standardize our test sample of SNe
Ia with an RMS of $0.091 \pm 0.010$ mag, which corresponds to $0.074 \pm 0.010$
mag if peculiar velocity contributions are removed. Trained models and codes
are released at
\href{https://github.com/georgestein/suPAErnova}{github.com/georgestein/suPAErnova}
Related papers
- Graph Spatiotemporal Process for Multivariate Time Series Anomaly
Detection with Missing Values [67.76168547245237]
We introduce a novel framework called GST-Pro, which utilizes a graphtemporal process and anomaly scorer to detect anomalies.
Our experimental results show that the GST-Pro method can effectively detect anomalies in time series data and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-01-11T10:10:16Z) - Generative Modeling of Regular and Irregular Time Series Data via Koopman VAEs [50.25683648762602]
We introduce Koopman VAE, a new generative framework that is based on a novel design for the model prior.
Inspired by Koopman theory, we represent the latent conditional prior dynamics using a linear map.
KoVAE outperforms state-of-the-art GAN and VAE methods across several challenging synthetic and real-world time series generation benchmarks.
arXiv Detail & Related papers (2023-10-04T07:14:43Z) - One-Dimensional Deep Image Prior for Curve Fitting of S-Parameters from
Electromagnetic Solvers [57.441926088870325]
Deep Image Prior (DIP) is a technique that optimized the weights of a randomly-d convolutional neural network to fit a signal from noisy or under-determined measurements.
Relative to publicly available implementations of Vector Fitting (VF), our method shows superior performance on nearly all test examples.
arXiv Detail & Related papers (2023-06-06T20:28:37Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - LSTM and CNN application for core-collapse supernova search in
gravitational wave real data [0.0]
Core-collapse supernovae (CCSNe) are expected to emit gravitational wave signals that could be detected by interferometers within the Milky Way and nearby galaxies.
We show potential of machine learning (ML) for multi-label classification of different CCSNe simulated signals and noise transients using real data.
arXiv Detail & Related papers (2023-01-23T12:12:33Z) - Imputing Missing Observations with Time Sliced Synthetic Minority
Oversampling Technique [0.3973560285628012]
We present a simple yet novel time series imputation technique with the goal of constructing an irregular time series that is uniform across every sample in a data set.
We fix a grid defined by the midpoints of non-overlapping bins (dubbed "slices") of observation times and ensure that each sample has values for all of the features at that given time.
This allows one to both impute fully missing observations to allow uniform time series classification across the entire data and, in special cases, to impute individually missing features.
arXiv Detail & Related papers (2022-01-14T19:23:24Z) - Emulating Spatio-Temporal Realizations of Three-Dimensional Isotropic
Turbulence via Deep Sequence Learning Models [24.025975236316842]
We use a data-driven approach to model a three-dimensional turbulent flow using cutting-edge Deep Learning techniques.
The accuracy of the model is assessed using statistical and physics-based metrics.
arXiv Detail & Related papers (2021-12-07T03:33:39Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled
Markov Chains [34.77971292478243]
The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture.
We develop a training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient.
We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.
arXiv Detail & Related papers (2020-10-05T08:11:55Z) - Gravitational-wave parameter estimation with autoregressive neural
network flows [0.0]
We introduce the use of autoregressive normalizing flows for rapid likelihood-free inference of binary black hole system parameters from gravitational-wave data with deep neural networks.
A normalizing flow is an invertible mapping on a sample space that can be used to induce a transformation from a simple probability distribution to a more complex one.
We build a more powerful latent variable model by incorporating autoregressive flows within the variational autoencoder framework.
arXiv Detail & Related papers (2020-02-18T15:44:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.