Gradient-based Bayesian Experimental Design for Implicit Models using
Mutual Information Lower Bounds
- URL: http://arxiv.org/abs/2105.04379v1
- Date: Mon, 10 May 2021 13:59:25 GMT
- Title: Gradient-based Bayesian Experimental Design for Implicit Models using
Mutual Information Lower Bounds
- Authors: Steven Kleinegesse and Michael U. Gutmann
- Abstract summary: We introduce a framework for Bayesian experimental design (BED) with implicit models, where the data-generating distribution is intractable but sampling from it is still possible.
In order to find optimal experimental designs for such models, our approach maximises mutual information lower bounds that are parametrised by neural networks.
By training a neural network on sampled data, we simultaneously update network parameters and designs using gradient-ascent.
- Score: 20.393359858407162
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a framework for Bayesian experimental design (BED) with implicit
models, where the data-generating distribution is intractable but sampling from
it is still possible. In order to find optimal experimental designs for such
models, our approach maximises mutual information lower bounds that are
parametrised by neural networks. By training a neural network on sampled data,
we simultaneously update network parameters and designs using stochastic
gradient-ascent. The framework enables experimental design with a variety of
prominent lower bounds and can be applied to a wide range of scientific tasks,
such as parameter estimation, model discrimination and improving future
predictions. Using a set of intractable toy models, we provide a comprehensive
empirical comparison of prominent lower bounds applied to the aforementioned
tasks. We further validate our framework on a challenging system of stochastic
differential equations from epidemiology.
Related papers
- Supervised Score-Based Modeling by Gradient Boosting [49.556736252628745]
We propose a Supervised Score-based Model (SSM) which can be viewed as a gradient boosting algorithm combining score matching.
We provide a theoretical analysis of learning and sampling for SSM to balance inference time and prediction accuracy.
Our model outperforms existing models in both accuracy and inference time.
arXiv Detail & Related papers (2024-11-02T07:06:53Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - Online simulator-based experimental design for cognitive model selection [74.76661199843284]
We propose BOSMOS: an approach to experimental design that can select between computational models without tractable likelihoods.
In simulated experiments, we demonstrate that the proposed BOSMOS technique can accurately select models in up to 2 orders of magnitude less time than existing LFI alternatives.
arXiv Detail & Related papers (2023-03-03T21:41:01Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Deep Variational Models for Collaborative Filtering-based Recommender
Systems [63.995130144110156]
Deep learning provides accurate collaborative filtering models to improve recommender system results.
Our proposed models apply the variational concept to injectity in the latent space of the deep architecture.
Results show the superiority of the proposed approach in scenarios where the variational enrichment exceeds the injected noise effect.
arXiv Detail & Related papers (2021-07-27T08:59:39Z) - Fully differentiable model discovery [0.0]
We propose an approach by combining neural network based surrogates with Sparse Bayesian Learning.
Our work expands PINNs to various types of neural network architectures, and connects neural network-based surrogates to the rich field of Bayesian parameter inference.
arXiv Detail & Related papers (2021-06-09T08:11:23Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Sequential Bayesian Experimental Design for Implicit Models via Mutual
Information [12.68659360172393]
A class of models of particular interest for the natural and medical sciences are implicit models.
We devise a novel sequential design framework for parameter estimation that uses the Mutual Information (MI) between model parameters and simulated data as a utility function.
We find that our framework is efficient for the various implicit models tested, yielding accurate parameter estimates after only a few iterations.
arXiv Detail & Related papers (2020-03-20T16:52:10Z) - Bayesian Experimental Design for Implicit Models by Mutual Information
Neural Estimation [16.844481439960663]
Implicit models, where the data-generation distribution is intractable but sampling is possible, are ubiquitous in the natural sciences.
A fundamental question is how to design experiments so that the collected data are most useful.
For implicit models, however, this approach is severely hampered by the high computational cost of computing posteriors.
We show that training a neural network to maximise a lower bound on MI allows us to jointly determine the optimal design and the posterior.
arXiv Detail & Related papers (2020-02-19T12:09:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.