Convolutional neural networks as an alternative to Bayesian retrievals
- URL: http://arxiv.org/abs/2203.01236v2
- Date: Thu, 3 Mar 2022 13:02:26 GMT
- Title: Convolutional neural networks as an alternative to Bayesian retrievals
- Authors: Francisco Ardevol Martinez, Michiel Min, Inga Kamp, Paul I. Palmer
- Abstract summary: We compare machine learning retrievals of exoplanet transmission spectra with nested sampling.
We also test how robust machine learning and nested sampling are against incorrect assumptions in our models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Exoplanet observations are currently analysed with Bayesian retrieval
techniques. Due to the computational load of the models used, a compromise is
needed between model complexity and computing time. Analysis of data from
future facilities, will need more complex models which will increase the
computational load of retrievals, prompting the search for a faster approach
for interpreting exoplanet observations. Our goal is to compare machine
learning retrievals of exoplanet transmission spectra with nested sampling, and
understand if machine learning can be as reliable as Bayesian retrievals for a
statistically significant sample of spectra while being orders of magnitude
faster. We generate grids of synthetic transmission spectra and their
corresponding planetary and atmospheric parameters, one using free chemistry
models, and the other using equilibrium chemistry models. Each grid is
subsequently rebinned to simulate both HST/WFC3 and JWST/NIRSpec observations,
yielding four datasets in total. Convolutional neural networks (CNNs) are
trained with each of the datasets. We perform retrievals on a 1,000 simulated
observations for each combination of model type and instrument with nested
sampling and machine learning. We also use both methods to perform retrievals
on real WFC3 transmission spectra. Finally, we test how robust machine learning
and nested sampling are against incorrect assumptions in our models. CNNs reach
a lower coefficient of determination between predicted and true values of the
parameters. Nested sampling underestimates the uncertainty in ~8% of
retrievals, whereas CNNs estimate them correctly. For real WFC3 observations,
nested sampling and machine learning agree within $2\sigma$ for ~86% of
spectra. When doing retrievals with incorrect assumptions, nested sampling
underestimates the uncertainty in ~12% to ~41% of cases, whereas this is always
below ~10% for the CNN.
Related papers
- Stochastic Approximation Approach to Federated Machine Learning [0.0]
This paper examines Federated learning (FL) in a Approximation (SA) framework.
FL is a collaborative way to train neural network models across various participants or clients.
It is observed that the proposed algorithm is robust and gives more reliable estimates of the weights.
arXiv Detail & Related papers (2024-02-20T12:00:25Z) - Assessing Neural Network Representations During Training Using
Noise-Resilient Diffusion Spectral Entropy [55.014926694758195]
Entropy and mutual information in neural networks provide rich information on the learning process.
We leverage data geometry to access the underlying manifold and reliably compute these information-theoretic measures.
We show that they form noise-resistant measures of intrinsic dimensionality and relationship strength in high-dimensional simulated data.
arXiv Detail & Related papers (2023-12-04T01:32:42Z) - Sampling weights of deep neural networks [1.2370077627846041]
We introduce a probability distribution, combined with an efficient sampling algorithm, for weights and biases of fully-connected neural networks.
In a supervised learning context, no iterative optimization or gradient computations of internal network parameters are needed.
We prove that sampled networks are universal approximators.
arXiv Detail & Related papers (2023-06-29T10:13:36Z) - Revisiting the Evaluation of Image Synthesis with GANs [55.72247435112475]
This study presents an empirical investigation into the evaluation of synthesis performance, with generative adversarial networks (GANs) as a representative of generative models.
In particular, we make in-depth analyses of various factors, including how to represent a data point in the representation space, how to calculate a fair distance using selected samples, and how many instances to use from each set.
arXiv Detail & Related papers (2023-04-04T17:54:32Z) - Automatic Neural Network Hyperparameter Optimization for Extrapolation:
Lessons Learned from Visible and Near-Infrared Spectroscopy of Mango Fruit [7.462336024223667]
This paper considers automatic methods for configuring a neural network that extrapolates in time for the domain of visible and near-infrared (VNIR) spectroscopy.
To encourage the neural network model to extrapolate, we consider validating model configurations on samples that are shifted in time similar to the test set.
We find that ensembling improves the state-of-the-art model's variance and accuracy.
arXiv Detail & Related papers (2022-10-03T00:41:05Z) - BeCAPTCHA-Type: Biometric Keystroke Data Generation for Improved Bot
Detection [63.447493500066045]
This work proposes a data driven learning model for the synthesis of keystroke biometric data.
The proposed method is compared with two statistical approaches based on Universal and User-dependent models.
Our experimental framework considers a dataset with 136 million keystroke events from 168 thousand subjects.
arXiv Detail & Related papers (2022-07-27T09:26:15Z) - A Probabilistic Autoencoder for Type Ia Supernovae Spectral Time Series [1.3316902142331577]
We construct a probabilistic autoencoder to learn the intrinsic diversity of type Ia supernovae (SNe Ia) from a sparse set of spectral time series.
We show that the PAE learns a low-dimensional latent space that captures the nonlinear range of features that exists within the population.
We then use our PAE in a number of downstream tasks on SNe Ia for increasingly precise cosmological analyses.
arXiv Detail & Related papers (2022-07-15T17:58:27Z) - Learning Summary Statistics for Bayesian Inference with Autoencoders [58.720142291102135]
We use the inner dimension of deep neural network based Autoencoders as summary statistics.
To create an incentive for the encoder to encode all the parameter-related information but not the noise, we give the decoder access to explicit or implicit information that has been used to generate the training data.
arXiv Detail & Related papers (2022-01-28T12:00:31Z) - Transfer Learning with Convolutional Networks for Atmospheric Parameter
Retrieval [14.131127382785973]
The Infrared Atmospheric Sounding Interferometer (IASI) on board the MetOp satellite series provides important measurements for Numerical Weather Prediction (NWP)
Retrieving accurate atmospheric parameters from the raw data provided by IASI is a large challenge, but necessary in order to use the data in NWP models.
We show how features extracted from the IASI data by a CNN trained to predict a physical variable can be used as inputs to another statistical method designed to predict a different physical variable at low altitude.
arXiv Detail & Related papers (2020-12-09T09:28:42Z) - Network Classifiers Based on Social Learning [71.86764107527812]
We propose a new way of combining independently trained classifiers over space and time.
The proposed architecture is able to improve prediction performance over time with unlabeled data.
We show that this strategy results in consistent learning with high probability, and it yields a robust structure against poorly trained classifiers.
arXiv Detail & Related papers (2020-10-23T11:18:20Z) - Parameter Space Factorization for Zero-Shot Learning across Tasks and
Languages [112.65994041398481]
We propose a Bayesian generative model for the space of neural parameters.
We infer the posteriors over such latent variables based on data from seen task-language combinations.
Our model yields comparable or better results than state-of-the-art, zero-shot cross-lingual transfer methods.
arXiv Detail & Related papers (2020-01-30T16:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.