A Bayesian machine scientist to aid in the solution of challenging
scientific problems
- URL: http://arxiv.org/abs/2004.12157v1
- Date: Sat, 25 Apr 2020 14:42:13 GMT
- Title: A Bayesian machine scientist to aid in the solution of challenging
scientific problems
- Authors: Roger Guimera and Ignasi Reichardt and Antoni Aguilar-Mogas and
Francesco A Massucci and Manuel Miranda and Jordi Pallares and Marta
Sales-Pardo
- Abstract summary: We introduce a Bayesian machine scientist, which establishes the plausibility of models using explicit approximations to the exact marginal posterior over models.
It explores the space of models using Markov chain Monte Carlo.
We show that this approach uncovers accurate models for synthetic and real data and provides out-of-sample predictions that are more accurate than those of existing approaches.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Closed-form, interpretable mathematical models have been instrumental for
advancing our understanding of the world; with the data revolution, we may now
be in a position to uncover new such models for many systems from physics to
the social sciences. However, to deal with increasing amounts of data, we need
"machine scientists" that are able to extract these models automatically from
data. Here, we introduce a Bayesian machine scientist, which establishes the
plausibility of models using explicit approximations to the exact marginal
posterior over models and establishes its prior expectations about models by
learning from a large empirical corpus of mathematical expressions. It explores
the space of models using Markov chain Monte Carlo. We show that this approach
uncovers accurate models for synthetic and real data and provides out-of-sample
predictions that are more accurate than those of existing approaches and of
other nonparametric methods.
Related papers
- Towards Learning Stochastic Population Models by Gradient Descent [0.0]
We show that simultaneous estimation of parameters and structure poses major challenges for optimization procedures.
We demonstrate accurate estimation of models but find that enforcing the inference of parsimonious, interpretable models drastically increases the difficulty.
arXiv Detail & Related papers (2024-04-10T14:38:58Z) - Fusion of Gaussian Processes Predictions with Monte Carlo Sampling [61.31380086717422]
In science and engineering, we often work with models designed for accurate prediction of variables of interest.
Recognizing that these models are approximations of reality, it becomes desirable to apply multiple models to the same data and integrate their outcomes.
arXiv Detail & Related papers (2024-03-03T04:21:21Z) - Discovering interpretable models of scientific image data with deep
learning [0.0]
We implement representation learning, sparse deep neural network training and symbolic regression.
We demonstrate their relevance to the field of bioimaging using a well-studied test problem of classifying cell states in microscopy data.
We explore the utility of such interpretable models in producing scientific explanations of the underlying biological phenomenon.
arXiv Detail & Related papers (2024-02-05T15:45:55Z) - Discovering Interpretable Physical Models using Symbolic Regression and
Discrete Exterior Calculus [55.2480439325792]
We propose a framework that combines Symbolic Regression (SR) and Discrete Exterior Calculus (DEC) for the automated discovery of physical models.
DEC provides building blocks for the discrete analogue of field theories, which are beyond the state-of-the-art applications of SR to physical problems.
We prove the effectiveness of our methodology by re-discovering three models of Continuum Physics from synthetic experimental data.
arXiv Detail & Related papers (2023-10-10T13:23:05Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - Breiman's two cultures: You don't have to choose sides [10.695407438192527]
Breiman's classic paper casts data analysis as a choice between two cultures.
Data modelers use simple, interpretable models with well-understood theoretical properties to analyze data.
Algorithm modelers prioritize predictive accuracy and use more flexible function approximations to analyze data.
arXiv Detail & Related papers (2021-04-25T17:58:46Z) - Kernel-Based Models for Influence Maximization on Graphs based on
Gaussian Process Variance Minimization [9.357483974291899]
We introduce and investigate a novel model for influence (IM) on graphs.
Data-driven approaches can be applied to determine proper kernels for this IM model.
Compared to models in this field that rely on costly Monte-Carlo simulations, our model allows for a simple and cost-efficient update strategy.
arXiv Detail & Related papers (2021-03-02T08:55:34Z) - Design of Dynamic Experiments for Black-Box Model Discrimination [72.2414939419588]
Consider a dynamic model discrimination setting where we wish to chose: (i) what is the best mechanistic, time-varying model and (ii) what are the best model parameter estimates.
For rival mechanistic models where we have access to gradient information, we extend existing methods to incorporate a wider range of problem uncertainty.
We replace these black-box models with Gaussian process surrogate models and thereby extend the model discrimination setting to additionally incorporate rival black-box model.
arXiv Detail & Related papers (2021-02-07T11:34:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.