A Likelihood-Free Approach to Goal-Oriented Bayesian Optimal Experimental Design
- URL: http://arxiv.org/abs/2408.09582v1
- Date: Sun, 18 Aug 2024 19:45:49 GMT
- Title: A Likelihood-Free Approach to Goal-Oriented Bayesian Optimal Experimental Design
- Authors: Atlanta Chakraborty, Xun Huan, Tommie Catanach,
- Abstract summary: We introduce LF-GO-OED (likelihood-free goal-oriented optimal experimental design), a computational method for conducting GO-OED with nonlinear observation and prediction models.
It is specifically designed to accommodate implicit models, where the likelihood is intractable.
The method is validated on benchmark problems with existing methods, and demonstrated on scientific applications of epidemiology and neural science.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conventional Bayesian optimal experimental design seeks to maximize the expected information gain (EIG) on model parameters. However, the end goal of the experiment often is not to learn the model parameters, but to predict downstream quantities of interest (QoIs) that depend on the learned parameters. And designs that offer high EIG for parameters may not translate to high EIG for QoIs. Goal-oriented optimal experimental design (GO-OED) thus directly targets to maximize the EIG of QoIs. We introduce LF-GO-OED (likelihood-free goal-oriented optimal experimental design), a computational method for conducting GO-OED with nonlinear observation and prediction models. LF-GO-OED is specifically designed to accommodate implicit models, where the likelihood is intractable. In particular, it builds a density ratio estimator from samples generated from approximate Bayesian computation (ABC), thereby sidestepping the need for likelihood evaluations or density estimations. The overall method is validated on benchmark problems with existing methods, and demonstrated on scientific applications of epidemiology and neural science.
Related papers
- Goal-Oriented Bayesian Optimal Experimental Design for Nonlinear Models using Markov Chain Monte Carlo [0.0]
We present a computational framework of predictive goal-oriented OED (GO-OED) suitable for nonlinear observation and prediction models.
GO-OED seeks the experimental design providing the greatest EIG on the QoIs.
We demonstrate the effectiveness of the overall nonlinear GO-OED method, and illustrate its differences versus conventional non-GO-OED.
arXiv Detail & Related papers (2024-03-26T19:49:58Z) - Protein Design with Guided Discrete Diffusion [67.06148688398677]
A popular approach to protein design is to combine a generative model with a discriminative model for conditional sampling.
We propose diffusioN Optimized Sampling (NOS), a guidance method for discrete diffusion models.
NOS makes it possible to perform design directly in sequence space, circumventing significant limitations of structure-based methods.
arXiv Detail & Related papers (2023-05-31T16:31:24Z) - Statistically Efficient Bayesian Sequential Experiment Design via
Reinforcement Learning with Cross-Entropy Estimators [15.461927416747582]
Reinforcement learning can learn amortised design policies for designing sequences of experiments.
We propose the use of an alternative estimator based on the cross-entropy of the joint model distribution and a flexible proposal distribution.
Our method overcomes the exponential-sample complexity of previous approaches and provide more accurate estimates of high EIG values.
arXiv Detail & Related papers (2023-05-29T00:35:52Z) - Prediction-Oriented Bayesian Active Learning [51.426960808684655]
Expected predictive information gain (EPIG) is an acquisition function that measures information gain in the space of predictions rather than parameters.
EPIG leads to stronger predictive performance compared with BALD across a range of datasets and models.
arXiv Detail & Related papers (2023-04-17T10:59:57Z) - Metrics for Bayesian Optimal Experiment Design under Model
Misspecification [3.04585143845864]
The utility function defines the objective of the experiment where a common utility function is the information gain.
This article introduces an expanded framework for this process, where we go beyond the traditional Expected Information Gain criteria.
The functionality of the framework is showcased through its application to a scenario involving a linearized spring mass damper system and an F-16 model.
arXiv Detail & Related papers (2023-04-17T02:13:20Z) - Validation Diagnostics for SBI algorithms based on Normalizing Flows [55.41644538483948]
This work proposes easy to interpret validation diagnostics for multi-dimensional conditional (posterior) density estimators based on NF.
It also offers theoretical guarantees based on results of local consistency.
This work should help the design of better specified models or drive the development of novel SBI-algorithms.
arXiv Detail & Related papers (2022-11-17T15:48:06Z) - Design Amortization for Bayesian Optimal Experimental Design [70.13948372218849]
We build off of successful variational approaches, which optimize a parameterized variational model with respect to bounds on the expected information gain (EIG)
We present a novel neural architecture that allows experimenters to optimize a single variational model that can estimate the EIG for potentially infinitely many designs.
arXiv Detail & Related papers (2022-10-07T02:12:34Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - SUMO: Unbiased Estimation of Log Marginal Probability for Latent
Variable Models [80.22609163316459]
We introduce an unbiased estimator of the log marginal likelihood and its gradients for latent variable models based on randomized truncation of infinite series.
We show that models trained using our estimator give better test-set likelihoods than a standard importance-sampling based approach for the same average computational cost.
arXiv Detail & Related papers (2020-04-01T11:49:30Z) - Bayesian Experimental Design for Implicit Models by Mutual Information
Neural Estimation [16.844481439960663]
Implicit models, where the data-generation distribution is intractable but sampling is possible, are ubiquitous in the natural sciences.
A fundamental question is how to design experiments so that the collected data are most useful.
For implicit models, however, this approach is severely hampered by the high computational cost of computing posteriors.
We show that training a neural network to maximise a lower bound on MI allows us to jointly determine the optimal design and the posterior.
arXiv Detail & Related papers (2020-02-19T12:09:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.