Savage-Dickey density ratio estimation with normalizing flows for Bayesian model comparison
- URL: http://arxiv.org/abs/2506.04339v1
- Date: Wed, 04 Jun 2025 18:00:24 GMT
- Title: Savage-Dickey density ratio estimation with normalizing flows for Bayesian model comparison
- Authors: Kiyam Lin, Alicja Polanska, Davide Piras, Alessio Spurio Mancini, Jason D. McEwen,
- Abstract summary: We use the Savage-Dickey density ratio to calculate the Bayes factor (evidence ratio) between two nested models.<n>We introduce a neural SDDR approach using normalizing flows that can scale to settings where the super model contains a large number of extra parameters.<n>For a field-level inference setting, we show that Bayes factors computed for a Bayesian hierarchical model and simulation-based inference ( SBI) approach are consistent.
- Score: 4.232577149837663
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A core motivation of science is to evaluate which scientific model best explains observed data. Bayesian model comparison provides a principled statistical approach to comparing scientific models and has found widespread application within cosmology and astrophysics. Calculating the Bayesian evidence is computationally challenging, especially as we continue to explore increasingly more complex models. The Savage-Dickey density ratio (SDDR) provides a method to calculate the Bayes factor (evidence ratio) between two nested models using only posterior samples from the super model. The SDDR requires the calculation of a normalised marginal distribution over the extra parameters of the super model, which has typically been performed using classical density estimators, such as histograms. Classical density estimators, however, can struggle to scale to high-dimensional settings. We introduce a neural SDDR approach using normalizing flows that can scale to settings where the super model contains a large number of extra parameters. We demonstrate the effectiveness of this neural SDDR methodology applied to both toy and realistic cosmological examples. For a field-level inference setting, we show that Bayes factors computed for a Bayesian hierarchical model (BHM) and simulation-based inference (SBI) approach are consistent, providing further validation that SBI extracts as much cosmological information from the field as the BHM approach. The SDDR estimator with normalizing flows is implemented in the open-source harmonic Python package.
Related papers
- Discriminative versus Generative Approaches to Simulation-based Inference [0.19999259391104385]
Deep learning has enabled unbinned and high-dimensional parameter estimation.<n>We compare two approaches for neural simulation-based inference (N SBI)<n>We find that both the direct likelihood and likelihood ratio estimation are able to effectively extract parameters with reasonable uncertainties.
arXiv Detail & Related papers (2025-03-11T01:38:54Z) - Bayesian Circular Regression with von Mises Quasi-Processes [57.88921637944379]
In this work we explore a family of expressive and interpretable distributions over circle-valued random functions.<n>For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Gibbs sampling.<n>We present experiments applying this model to the prediction of wind directions and the percentage of the running gait cycle as a function of joint angles.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - Consistent and fast inference in compartmental models of epidemics using
Poisson Approximate Likelihoods [1.933681537640272]
We introduce Poisson Approximate Likelihood (PAL) methods for epidemiological inference.
PALs are simple to implement, involving only elementary arithmetic operations and no tuning parameters.
We show how PALs can be used to: fit an age-structured model of influenza, taking advantage of automatic differentiation in Stan; compare over-dispersion in rotavirus by embedding PALs within sequential Monte Carlo.
arXiv Detail & Related papers (2022-05-26T20:19:28Z) - Model Comparison in Approximate Bayesian Computation [0.456877715768796]
A common problem in natural sciences is the comparison of competing models in the light of observed data.
This framework relies on the calculation of likelihood functions which are intractable for most models used in practice.
I propose a new efficient method to perform Bayesian model comparison in ABC.
arXiv Detail & Related papers (2022-03-15T10:24:16Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z) - Post-mortem on a deep learning contest: a Simpson's paradox and the
complementary roles of scale metrics versus shape metrics [61.49826776409194]
We analyze a corpus of models made publicly-available for a contest to predict the generalization accuracy of neural network (NN) models.
We identify what amounts to a Simpson's paradox: where "scale" metrics perform well overall but perform poorly on sub partitions of the data.
We present two novel shape metrics, one data-independent, and the other data-dependent, which can predict trends in the test accuracy of a series of NNs.
arXiv Detail & Related papers (2021-06-01T19:19:49Z) - Approximate Bayesian inference from noisy likelihoods with Gaussian
process emulated MCMC [0.24275655667345403]
We model the log-likelihood function using a Gaussian process (GP)
The main methodological innovation is to apply this model to emulate the progression that an exact Metropolis-Hastings (MH) sampler would take.
The resulting approximate sampler is conceptually simple and sample-efficient.
arXiv Detail & Related papers (2021-04-08T17:38:02Z) - Leveraging Global Parameters for Flow-based Neural Posterior Estimation [90.21090932619695]
Inferring the parameters of a model based on experimental observations is central to the scientific method.
A particularly challenging setting is when the model is strongly indeterminate, i.e., when distinct sets of parameters yield identical observations.
We present a method for cracking such indeterminacy by exploiting additional information conveyed by an auxiliary set of observations sharing global parameters.
arXiv Detail & Related papers (2021-02-12T12:23:13Z) - Referenced Thermodynamic Integration for Bayesian Model Selection:
Application to COVID-19 Model Selection [1.9599274203282302]
We show how to compute the ratio of two models' normalising constants, known as the Bayes factor.
In this paper we apply a variation of the TI method, referred to as referenced TI, which computes a single model's normalising constant in an efficient way.
The approach is shown to be useful in practice when applied to a real problem - to perform model selection for a semi-mechanistic hierarchical Bayesian model of COVID-19 transmission in South Korea.
arXiv Detail & Related papers (2020-09-08T16:32:06Z) - An Empirical Comparison of GANs and Normalizing Flows for Density
Estimation [5.837881923712393]
Generative adversarial networks (GANs) and normalizing flows are approaches to density estimation that use deep neural networks.
GANs and normalizing flows have seldom been compared to each other for modeling non-image data.
No GAN is capable of modeling our simple low-dimensional data well, a task we view as a prerequisite for an approach to be considered suitable for general-purpose statistical modeling.
arXiv Detail & Related papers (2020-06-17T21:56:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.