Exoplanet Characterization using Conditional Invertible Neural Networks
- URL: http://arxiv.org/abs/2202.00027v1
- Date: Mon, 31 Jan 2022 19:00:06 GMT
- Title: Exoplanet Characterization using Conditional Invertible Neural Networks
- Authors: Jonas Haldemann, Victor Ksoll, Daniel Walter, Yann Alibert, Ralf S.
Klessen, Willy Benz, Ullrich Koethe, Lynton Ardizzone, Carsten Rother
- Abstract summary: Conditional invertible neural networks (cINNs) are a special type of neural network which excel in solving inverse problems.
We show that cINNs are a possible alternative to the standard time-consuming sampling methods.
- Score: 21.516242058639637
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The characterization of an exoplanet's interior is an inverse problem, which
requires statistical methods such as Bayesian inference in order to be solved.
Current methods employ Markov Chain Monte Carlo (MCMC) sampling to infer the
posterior probability of planetary structure parameters for a given exoplanet.
These methods are time consuming since they require the calculation of a large
number of planetary structure models. To speed up the inference process when
characterizing an exoplanet, we propose to use conditional invertible neural
networks (cINNs) to calculate the posterior probability of the internal
structure parameters. cINNs are a special type of neural network which excel in
solving inverse problems. We constructed a cINN using FrEIA, which was then
trained on a database of $5.6\cdot 10^6$ internal structure models to recover
the inverse mapping between internal structure parameters and observable
features (i.e., planetary mass, planetary radius and composition of the host
star). The cINN method was compared to a Metropolis-Hastings MCMC. For that we
repeated the characterization of the exoplanet K2-111 b, using both the MCMC
method and the trained cINN. We show that the inferred posterior probability of
the internal structure parameters from both methods are very similar, with the
biggest differences seen in the exoplanet's water content. Thus cINNs are a
possible alternative to the standard time-consuming sampling methods. Indeed,
using cINNs allows for orders of magnitude faster inference of an exoplanet's
composition than what is possible using an MCMC method, however, it still
requires the computation of a large database of internal structures to train
the cINN. Since this database is only computed once, we found that using a cINN
is more efficient than an MCMC, when more than 10 exoplanets are characterized
using the same cINN.
Related papers
- NeuralCMS: A deep learning approach to study Jupiter's interior [0.0]
We propose an efficient deep neural network (DNN) model to generate high-precision wide-ranged interior models.
We trained a sharing-based DNN with a large set of CMS results for a four-layer interior model of Jupiter.
NeuralCMS shows very good performance in predicting the gravity moments, with errors comparable with the uncertainty due to differential rotation, and a very accurate mass prediction.
arXiv Detail & Related papers (2024-05-15T10:55:16Z) - Universal Neural Functionals [67.80283995795985]
A challenging problem in many modern machine learning tasks is to process weight-space features.
Recent works have developed promising weight-space models that are equivariant to the permutation symmetries of simple feedforward networks.
This work proposes an algorithm that automatically constructs permutation equivariant models for any weight space.
arXiv Detail & Related papers (2024-02-07T20:12:27Z) - FlopPITy: Enabling self-consistent exoplanet atmospheric retrievals with
machine learning [0.0]
We implement and test sequential neural posterior estimation (SNPE) for exoplanet atmospheric retrievals.
The goal is to speed up retrievals so they can be run with more computationally expensive atmospheric models.
We generate 100 synthetic observations using ARCiS and perform retrievals on them to test the faithfulness of the SNPE posteriors.
arXiv Detail & Related papers (2024-01-08T19:00:02Z) - ExoMDN: Rapid characterization of exoplanet interior structures with
Mixture Density Networks [0.0]
We present ExoMDN, a machine-learning model for the interior characterization of exoplanets.
We show that ExoMDN can deliver a full posterior distribution of mass fractions and thicknesses of each planetary layer in under a second on a standard Intel i5 CPU.
We use ExoMDN to characterize the interior of 22 confirmed exoplanets with mass and radius uncertainties below 10% and 5% respectively.
arXiv Detail & Related papers (2023-06-15T10:00:03Z) - A Recursively Recurrent Neural Network (R2N2) Architecture for Learning
Iterative Algorithms [64.3064050603721]
We generalize Runge-Kutta neural network to a recurrent neural network (R2N2) superstructure for the design of customized iterative algorithms.
We demonstrate that regular training of the weight parameters inside the proposed superstructure on input/output data of various computational problem classes yields similar iterations to Krylov solvers for linear equation systems, Newton-Krylov solvers for nonlinear equation systems, and Runge-Kutta solvers for ordinary differential equations.
arXiv Detail & Related papers (2022-11-22T16:30:33Z) - Differentiable and Transportable Structure Learning [73.84540901950616]
We introduce D-Struct, which recovers transportability in the discovered structures through a novel architecture and loss function.
Because D-Struct remains differentiable, our method can be easily adopted in existing differentiable architectures.
arXiv Detail & Related papers (2022-06-13T17:50:53Z) - Bayesian Structure Learning with Generative Flow Networks [85.84396514570373]
In Bayesian structure learning, we are interested in inferring a distribution over the directed acyclic graph (DAG) from data.
Recently, a class of probabilistic models, called Generative Flow Networks (GFlowNets), have been introduced as a general framework for generative modeling.
We show that our approach, called DAG-GFlowNet, provides an accurate approximation of the posterior over DAGs.
arXiv Detail & Related papers (2022-02-28T15:53:10Z) - Using Bayesian Deep Learning to infer Planet Mass from Gaps in
Protoplanetary Disks [0.0]
We introduce a deep learning network "DPNNet-Bayesian" that can predict planet mass from disk gaps.
A unique feature of our approach is that it can distinguish between the uncertainty associated with the deep learning architecture and uncertainty inherent in the input data.
The network predicts masses of $ 86.0 pm 5.5 M_Earth $, $ 43.8 pm 3.3 M_Earth $, and $ 92.2 pm 5.1 M_Earth $ respectively.
arXiv Detail & Related papers (2022-02-23T19:00:05Z) - PGNets: Planet mass prediction using convolutional neural networks for
radio continuum observations of protoplanetary disks [0.0]
Substructures induced by young planets in protoplanetary disks can be used to infer potential young planets' properties.
We developed Planet Gap neural Networks (PGNets) to infer planet mass from 2D images.
We reproduce the degeneracy scaling $alpha$ $propto $M_p3$ found in the linear fitting method.
arXiv Detail & Related papers (2021-11-30T08:12:08Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.