Bayesian Inference for the Multinomial Probit Model under Gaussian Prior
Distribution
- URL: http://arxiv.org/abs/2206.00720v1
- Date: Wed, 1 Jun 2022 19:10:41 GMT
- Title: Bayesian Inference for the Multinomial Probit Model under Gaussian Prior
Distribution
- Authors: Augusto Fasano, Giovanni Rebaudo, Niccol\`o Anceschi
- Abstract summary: Multinomial probit (mnp) models are fundamental and widely-applied regression models for categorical data.
Fasano and Durante (2022) proved that the class of unified skew-normal distributions is conjugate to several mnp sampling models.
We adapt the results for a popular special case: the discrete-choice mnp model under zero mean and independent Gaussian priors.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multinomial probit (mnp) models are fundamental and widely-applied regression
models for categorical data. Fasano and Durante (2022) proved that the class of
unified skew-normal distributions is conjugate to several mnp sampling models.
This allows to develop Monte Carlo samplers and accurate variational methods to
perform Bayesian inference. In this paper, we adapt the abovementioned results
for a popular special case: the discrete-choice mnp model under zero mean and
independent Gaussian priors. This allows to obtain simplified expressions for
the parameters of the posterior distribution and an alternative derivation for
the variational algorithm that gives a novel understanding of the fundamental
results in Fasano and Durante (2022) as well as computational advantages in our
special settings.
Related papers
- von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Fusion of Gaussian Processes Predictions with Monte Carlo Sampling [61.31380086717422]
In science and engineering, we often work with models designed for accurate prediction of variables of interest.
Recognizing that these models are approximations of reality, it becomes desirable to apply multiple models to the same data and integrate their outcomes.
arXiv Detail & Related papers (2024-03-03T04:21:21Z) - Diffusion models for probabilistic programming [56.47577824219207]
Diffusion Model Variational Inference (DMVI) is a novel method for automated approximate inference in probabilistic programming languages (PPLs)
DMVI is easy to implement, allows hassle-free inference in PPLs without the drawbacks of, e.g., variational inference using normalizing flows, and does not make any constraints on the underlying neural network model.
arXiv Detail & Related papers (2023-11-01T12:17:05Z) - Bayesian Neural Network Inference via Implicit Models and the Posterior
Predictive Distribution [0.8122270502556371]
We propose a novel approach to perform approximate Bayesian inference in complex models such as Bayesian neural networks.
The approach is more scalable to large data than Markov Chain Monte Carlo.
We see this being useful in applications such as surrogate and physics-based models.
arXiv Detail & Related papers (2022-09-06T02:43:19Z) - B\'ezier Curve Gaussian Processes [8.11969931278838]
This paper proposes a new probabilistic sequence model building on probabilistic B'ezier curves.
Combined with a Mixture Density network, Bayesian conditional inference can be performed without the need for mean field variational approximation.
The model is used for pedestrian trajectory prediction, where a generated prediction also serves as a GP prior.
arXiv Detail & Related papers (2022-05-03T19:49:57Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - Flexible mean field variational inference using mixtures of
non-overlapping exponential families [6.599344783327053]
I show that using standard mean field variational inference can fail to produce sensible results for models with sparsity-inducing priors.
I show that any mixture of a diffuse exponential family and a point mass at zero to model sparsity forms an exponential family.
arXiv Detail & Related papers (2020-10-14T01:46:56Z) - A Class of Conjugate Priors for Multinomial Probit Models which Includes
the Multivariate Normal One [0.3553493344868413]
We show that the entire class of unified skew-normal (SUN) distributions is conjugate to several multinomial probit models.
We improve upon state-of-the-art solutions for posterior inference and classification.
arXiv Detail & Related papers (2020-07-14T10:08:23Z) - Gaussian Process Regression with Local Explanation [28.90948136731314]
We propose GPR with local explanation, which reveals the feature contributions to the prediction of each sample.
In the proposed model, both the prediction and explanation for each sample are performed using an easy-to-interpret locally linear model.
For a new test sample, the proposed model can predict the values of its target variable and weight vector, as well as their uncertainties.
arXiv Detail & Related papers (2020-07-03T13:22:24Z) - Decision-Making with Auto-Encoding Variational Bayes [71.44735417472043]
We show that a posterior approximation distinct from the variational distribution should be used for making decisions.
Motivated by these theoretical results, we propose learning several approximate proposals for the best model.
In addition to toy examples, we present a full-fledged case study of single-cell RNA sequencing.
arXiv Detail & Related papers (2020-02-17T19:23:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.