Bayesian Model Selection via Mean-Field Variational Approximation
- URL: http://arxiv.org/abs/2312.10607v1
- Date: Sun, 17 Dec 2023 04:48:25 GMT
- Title: Bayesian Model Selection via Mean-Field Variational Approximation
- Authors: Yangfan Zhang, Yun Yang
- Abstract summary: We study the non-asymptotic properties of mean-field (MF) inference under the Bayesian framework.
We show a Bernstein von-Mises (BvM) theorem for the variational distribution from MF under possible model mis-specification.
- Score: 10.433170683584994
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article considers Bayesian model selection via mean-field (MF)
variational approximation. Towards this goal, we study the non-asymptotic
properties of MF inference under the Bayesian framework that allows latent
variables and model mis-specification. Concretely, we show a Bernstein
von-Mises (BvM) theorem for the variational distribution from MF under possible
model mis-specification, which implies the distributional convergence of MF
variational approximation to a normal distribution centering at the maximal
likelihood estimator (within the specified model). Motivated by the BvM
theorem, we propose a model selection criterion using the evidence lower bound
(ELBO), and demonstrate that the model selected by ELBO tends to asymptotically
agree with the one selected by the commonly used Bayesian information criterion
(BIC) as sample size tends to infinity. Comparing to BIC, ELBO tends to incur
smaller approximation error to the log-marginal likelihood (a.k.a. model
evidence) due to a better dimension dependence and full incorporation of the
prior information. Moreover, we show the geometric convergence of the
coordinate ascent variational inference (CAVI) algorithm under the parametric
model framework, which provides a practical guidance on how many iterations one
typically needs to run when approximating the ELBO. These findings demonstrate
that variational inference is capable of providing a computationally efficient
alternative to conventional approaches in tasks beyond obtaining point
estimates, which is also empirically demonstrated by our extensive numerical
experiments.
Related papers
- Stochastic Sampling from Deterministic Flow Models [8.849981177332594]
We present a method to turn flow models into a family of differential equations (SDEs) that have the same marginal distributions.
We empirically demonstrate advantages of our method on a toy Gaussian setup and on the large scale ImageNet generation task.
arXiv Detail & Related papers (2024-10-03T05:18:28Z) - Flow matching achieves almost minimax optimal convergence [50.38891696297888]
Flow matching (FM) has gained significant attention as a simulation-free generative model.
This paper discusses the convergence properties of FM for large sample size under the $p$-Wasserstein distance.
We establish that FM can achieve an almost minimax optimal convergence rate for $1 leq p leq 2$, presenting the first theoretical evidence that FM can reach convergence rates comparable to those of diffusion models.
arXiv Detail & Related papers (2024-05-31T14:54:51Z) - Estimating the Number of Components in Finite Mixture Models via Variational Approximation [8.468023518807408]
We introduce a new method for selecting the number of components in finite mixture models (FMMs) using variational Bayes.
We establish matching upper and lower bounds for the Evidence Lower Bound (ELBO) derived from mean-field (MF) variational approximation.
As a by-product of our proof, we demonstrate that the MF approximation inherits the stable behavior (benefited from model singularity) of the posterior distribution.
arXiv Detail & Related papers (2024-04-25T17:00:24Z) - Diffusion models for probabilistic programming [56.47577824219207]
Diffusion Model Variational Inference (DMVI) is a novel method for automated approximate inference in probabilistic programming languages (PPLs)
DMVI is easy to implement, allows hassle-free inference in PPLs without the drawbacks of, e.g., variational inference using normalizing flows, and does not make any constraints on the underlying neural network model.
arXiv Detail & Related papers (2023-11-01T12:17:05Z) - Bivariate Causal Discovery using Bayesian Model Selection [11.726586969589]
We show how to incorporate causal assumptions within the Bayesian framework.
This enables us to construct models with realistic assumptions.
We then outperform previous methods on a wide range of benchmark datasets.
arXiv Detail & Related papers (2023-06-05T14:51:05Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z) - Referenced Thermodynamic Integration for Bayesian Model Selection:
Application to COVID-19 Model Selection [1.9599274203282302]
We show how to compute the ratio of two models' normalising constants, known as the Bayes factor.
In this paper we apply a variation of the TI method, referred to as referenced TI, which computes a single model's normalising constant in an efficient way.
The approach is shown to be useful in practice when applied to a real problem - to perform model selection for a semi-mechanistic hierarchical Bayesian model of COVID-19 transmission in South Korea.
arXiv Detail & Related papers (2020-09-08T16:32:06Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Decision-Making with Auto-Encoding Variational Bayes [71.44735417472043]
We show that a posterior approximation distinct from the variational distribution should be used for making decisions.
Motivated by these theoretical results, we propose learning several approximate proposals for the best model.
In addition to toy examples, we present a full-fledged case study of single-cell RNA sequencing.
arXiv Detail & Related papers (2020-02-17T19:23:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.