A survey on Bayesian inference for Gaussian mixture model
- URL: http://arxiv.org/abs/2108.11753v1
- Date: Fri, 20 Aug 2021 13:23:17 GMT
- Title: A survey on Bayesian inference for Gaussian mixture model
- Authors: Jun Lu
- Abstract summary: The aim of this survey is to give a self-contained introduction to concepts and mathematical tools in Bayesian inference for finite and infinite Gaussian mixture model.
Other than this modest background, the development is self-contained, with rigorous proofs provided throughout.
- Score: 3.109306676759862
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Clustering has become a core technology in machine learning, largely due to
its application in the field of unsupervised learning, clustering,
classification, and density estimation. A frequentist approach exists to hand
clustering based on mixture model which is known as the EM algorithm where the
parameters of the mixture model are usually estimated into a maximum likelihood
estimation framework. Bayesian approach for finite and infinite Gaussian
mixture model generates point estimates for all variables as well as associated
uncertainty in the form of the whole estimates' posterior distribution.
The sole aim of this survey is to give a self-contained introduction to
concepts and mathematical tools in Bayesian inference for finite and infinite
Gaussian mixture model in order to seamlessly introduce their applications in
subsequent sections. However, we clearly realize our inability to cover all the
useful and interesting results concerning this field and given the paucity of
scope to present this discussion, e.g., the separated analysis of the
generation of Dirichlet samples by stick-breaking and Polya's Urn approaches.
We refer the reader to literature in the field of the Dirichlet process mixture
model for a much detailed introduction to the related fields. Some excellent
examples include (Frigyik et al., 2010; Murphy, 2012; Gelman et al., 2014;
Hoff, 2009).
This survey is primarily a summary of purpose, significance of important
background and techniques for Gaussian mixture model, e.g., Dirichlet prior,
Chinese restaurant process, and most importantly the origin and complexity of
the methods which shed light on their modern applications. The mathematical
prerequisite is a first course in probability. Other than this modest
background, the development is self-contained, with rigorous proofs provided
throughout.
Related papers
- von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Fusion of Gaussian Processes Predictions with Monte Carlo Sampling [61.31380086717422]
In science and engineering, we often work with models designed for accurate prediction of variables of interest.
Recognizing that these models are approximations of reality, it becomes desirable to apply multiple models to the same data and integrate their outcomes.
arXiv Detail & Related papers (2024-03-03T04:21:21Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - Bregman Power k-Means for Clustering Exponential Family Data [11.434503492579477]
We bridge algorithmic advances to classical work on hard clustering under Bregman divergences.
The elegant properties of Bregman divergences allow us to maintain closed form updates in a simple and transparent algorithm.
We consider thorough empirical analyses on simulated experiments and a case study on rainfall data, finding that the proposed method outperforms existing peer methods in a variety of non-Gaussian data settings.
arXiv Detail & Related papers (2022-06-22T06:09:54Z) - A Robust and Flexible EM Algorithm for Mixtures of Elliptical
Distributions with Missing Data [71.9573352891936]
This paper tackles the problem of missing data imputation for noisy and non-Gaussian data.
A new EM algorithm is investigated for mixtures of elliptical distributions with the property of handling potential missing data.
Experimental results on synthetic data demonstrate that the proposed algorithm is robust to outliers and can be used with non-Gaussian data.
arXiv Detail & Related papers (2022-01-28T10:01:37Z) - Probabilistic Mass Mapping with Neural Score Estimation [4.079848600120986]
We introduce a novel methodology for efficient sampling of the high-dimensional Bayesian posterior of the weak lensing mass-mapping problem.
We aim to demonstrate the accuracy of the method on simulations, and then proceed to applying it to the mass reconstruction of the HST/ACS COSMOS field.
arXiv Detail & Related papers (2022-01-14T17:07:48Z) - A Riemannian Newton Trust-Region Method for Fitting Gaussian Mixture
Models [0.0]
We introduce a formula for the Riemannian Hessian for Gaussian Mixture Models.
On top, we propose a new Newton Trust-Region method which outperforms current approaches both in terms of runtime and number of iterations.
arXiv Detail & Related papers (2021-04-30T12:48:32Z) - Pathwise Conditioning of Gaussian Processes [72.61885354624604]
Conventional approaches for simulating Gaussian process posteriors view samples as draws from marginal distributions of process values at finite sets of input locations.
This distribution-centric characterization leads to generative strategies that scale cubically in the size of the desired random vector.
We show how this pathwise interpretation of conditioning gives rise to a general family of approximations that lend themselves to efficiently sampling Gaussian process posteriors.
arXiv Detail & Related papers (2020-11-08T17:09:37Z) - Mean-Field Approximation to Gaussian-Softmax Integral with Application
to Uncertainty Estimation [23.38076756988258]
We propose a new single-model based approach to quantify uncertainty in deep neural networks.
We use a mean-field approximation formula to compute an analytically intractable integral.
Empirically, the proposed approach performs competitively when compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-06-13T07:32:38Z) - Nearest Neighbor Dirichlet Mixtures [3.3194866396158]
We propose a class of nearest neighbor-Dirichlet mixtures to maintain most of the strengths of Bayesian approaches without the computational disadvantages.
A simple and embarrassingly parallel Monte Carlo algorithm is proposed to sample from the resulting pseudo-posterior for the unknown density.
arXiv Detail & Related papers (2020-03-17T21:39:11Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.