Simple lessons from complex learning: what a neural network model learns
about cosmic structure formation
- URL: http://arxiv.org/abs/2206.04573v1
- Date: Thu, 9 Jun 2022 15:41:09 GMT
- Title: Simple lessons from complex learning: what a neural network model learns
about cosmic structure formation
- Authors: Drew Jamieson, Yin Li, Siyu He, Francisco Villaescusa-Navarro, Shirley
Ho, Renan Alves de Oliveira, David N. Spergel
- Abstract summary: We train a neural network model to predict the full phase space evolution of cosmological N-body simulations.
Our model achieves percent level accuracy at nonlinear scales of $ksim 1 mathrmMpc-1, h$, representing a significant improvement over COLA.
- Score: 7.270598539996841
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We train a neural network model to predict the full phase space evolution of
cosmological N-body simulations. Its success implies that the neural network
model is accurately approximating the Green's function expansion that relates
the initial conditions of the simulations to its outcome at later times in the
deeply nonlinear regime. We test the accuracy of this approximation by
assessing its performance on well understood simple cases that have either
known exact solutions or well understood expansions. These scenarios include
spherical configurations, isolated plane waves, and two interacting plane
waves: initial conditions that are very different from the Gaussian random
fields used for training. We find our model generalizes well to these well
understood scenarios, demonstrating that the networks have inferred general
physical principles and learned the nonlinear mode couplings from the complex,
random Gaussian training data. These tests also provide a useful diagnostic for
finding the model's strengths and weaknesses, and identifying strategies for
model improvement. We also test the model on initial conditions that contain
only transverse modes, a family of modes that differ not only in their phases
but also in their evolution from the longitudinal growing modes used in the
training set. When the network encounters these initial conditions that are
orthogonal to the training set, the model fails completely. In addition to
these simple configurations, we evaluate the model's predictions for the
density, displacement, and momentum power spectra with standard initial
conditions for N-body simulations. We compare these summary statistics against
N-body results and an approximate, fast simulation method called COLA. Our
model achieves percent level accuracy at nonlinear scales of $k\sim 1\
\mathrm{Mpc}^{-1}\, h$, representing a significant improvement over COLA.
Related papers
- The Convex Landscape of Neural Networks: Characterizing Global Optima
and Stationary Points via Lasso Models [75.33431791218302]
Deep Neural Network Network (DNN) models are used for programming purposes.
In this paper we examine the use of convex neural recovery models.
We show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
We also show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
arXiv Detail & Related papers (2023-12-19T23:04:56Z) - Learning Neural Constitutive Laws From Motion Observations for
Generalizable PDE Dynamics [97.38308257547186]
Many NN approaches learn an end-to-end model that implicitly models both the governing PDE and material models.
We argue that the governing PDEs are often well-known and should be explicitly enforced rather than learned.
We introduce a new framework termed "Neural Constitutive Laws" (NCLaw) which utilizes a network architecture that strictly guarantees standard priors.
arXiv Detail & Related papers (2023-04-27T17:42:24Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - NeuralFMU: Presenting a workflow for integrating hybrid NeuralODEs into
real world applications [0.0]
We present an intuitive workflow to setup and use NeuralFMUs.
We exemplify this concept by deploying a NeuralFMU for a consumption simulation based on a Vehicle Longitudinal Dynamics Model.
arXiv Detail & Related papers (2022-09-08T17:17:46Z) - Instability and Local Minima in GAN Training with Kernel Discriminators [20.362912591032636]
Generative Adversarial Networks (GANs) are a widely-used tool for generative modeling of complex data.
Despite their empirical success, the training of GANs is not fully understood due to the min-max optimization of the generator and discriminator.
This paper analyzes these joint dynamics when the true samples, as well as the generated samples, are discrete, finite sets, and the discriminator is kernel-based.
arXiv Detail & Related papers (2022-08-21T18:03:06Z) - An advanced spatio-temporal convolutional recurrent neural network for
storm surge predictions [73.4962254843935]
We study the capability of artificial neural network models to emulate storm surge based on the storm track/size/intensity history.
This study presents a neural network model that can predict storm surge, informed by a database of synthetic storm simulations.
arXiv Detail & Related papers (2022-04-18T23:42:18Z) - Bayesian Neural Ordinary Differential Equations [0.9422623204346027]
We demonstrate the successful integration of Neural ODEs with Bayesian inference frameworks.
We achieve a posterior sample accuracy of 98.5% on the test ensemble of 10,000 images.
This gives a scientific machine learning tool for probabilistic estimation of uncertainties.
arXiv Detail & Related papers (2020-12-14T04:05:26Z) - A Bayesian Perspective on Training Speed and Model Selection [51.15664724311443]
We show that a measure of a model's training speed can be used to estimate its marginal likelihood.
We verify our results in model selection tasks for linear models and for the infinite-width limit of deep neural networks.
Our results suggest a promising new direction towards explaining why neural networks trained with gradient descent are biased towards functions that generalize well.
arXiv Detail & Related papers (2020-10-27T17:56:14Z) - The Gaussian equivalence of generative models for learning with shallow
neural networks [30.47878306277163]
We study the performance of neural networks trained on data drawn from pre-trained generative models.
We provide three strands of rigorous, analytical and numerical evidence corroborating this equivalence.
These results open a viable path to the theoretical study of machine learning models with realistic data.
arXiv Detail & Related papers (2020-06-25T21:20:09Z) - Kernel and Rich Regimes in Overparametrized Models [69.40899443842443]
We show that gradient descent on overparametrized multilayer networks can induce rich implicit biases that are not RKHS norms.
We also demonstrate this transition empirically for more complex matrix factorization models and multilayer non-linear networks.
arXiv Detail & Related papers (2020-02-20T15:43:02Z) - Learning CHARME models with neural networks [1.5362025549031046]
We consider a model called CHARME (Conditional Heteroscedastic Autoregressive Mixture of Experts)
As an application, we develop a learning theory for the NN-based autoregressive functions of the model.
arXiv Detail & Related papers (2020-02-08T21:51:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.