A Riemannian Newton Trust-Region Method for Fitting Gaussian Mixture
Models
- URL: http://arxiv.org/abs/2104.14957v1
- Date: Fri, 30 Apr 2021 12:48:32 GMT
- Title: A Riemannian Newton Trust-Region Method for Fitting Gaussian Mixture
Models
- Authors: Lena Sembach, Jan Pablo Burgard, Volker H. Schulz
- Abstract summary: We introduce a formula for the Riemannian Hessian for Gaussian Mixture Models.
On top, we propose a new Newton Trust-Region method which outperforms current approaches both in terms of runtime and number of iterations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Gaussian Mixture Models are a powerful tool in Data Science and Statistics
that are mainly used for clustering and density approximation. The task of
estimating the model parameters is in practice often solved by the Expectation
Maximization (EM) algorithm which has its benefits in its simplicity and low
per-iteration costs. However, the EM converges slowly if there is a large share
of hidden information or overlapping clusters. Recent advances in Manifold
Optimization for Gaussian Mixture Models have gained increasing interest. We
introduce a formula for the Riemannian Hessian for Gaussian Mixture Models. On
top, we propose a new Riemannian Newton Trust-Region method which outperforms
current approaches both in terms of runtime and number of iterations.
Related papers
- von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Learning Mixtures of Gaussians Using Diffusion Models [9.118706387430883]
We give a new algorithm for learning mixtures of $k$ Gaussians to TV error.
Our approach is analytic and relies on the framework of diffusion models.
arXiv Detail & Related papers (2024-04-29T17:00:20Z) - Learning Gaussian Mixtures Using the Wasserstein-Fisher-Rao Gradient
Flow [12.455057637445174]
We propose a new algorithm to compute the nonparametric maximum likelihood estimator (NPMLE) in a Gaussian mixture model.
Our method is based on gradient descent over the space of probability measures equipped with the Wasserstein-Fisher-Rao geometry.
We conduct extensive numerical experiments to confirm the effectiveness of the proposed algorithm.
arXiv Detail & Related papers (2023-01-04T18:59:35Z) - Geometric Learning of Hidden Markov Models via a Method of Moments
Algorithm [11.338112397748619]
We present a novel algorithm for learning the parameters of hidden Markov models (HMMs) in a geometric setting.
We demonstrate through examples that the learner can result in significantly improved speed and numerical accuracy compared to existing learners.
arXiv Detail & Related papers (2022-07-02T12:24:38Z) - The Schr\"odinger Bridge between Gaussian Measures has a Closed Form [101.79851806388699]
We focus on the dynamic formulation of OT, also known as the Schr"odinger bridge (SB) problem.
In this paper, we provide closed-form expressions for SBs between Gaussian measures.
arXiv Detail & Related papers (2022-02-11T15:59:01Z) - Riemannian Score-Based Generative Modeling [56.20669989459281]
We introduce score-based generative models (SGMs) demonstrating remarkable empirical performance.
Current SGMs make the underlying assumption that the data is supported on a Euclidean manifold with flat geometry.
This prevents the use of these models for applications in robotics, geoscience or protein modeling.
arXiv Detail & Related papers (2022-02-06T11:57:39Z) - Bayesian Quadrature on Riemannian Data Manifolds [79.71142807798284]
A principled way to model nonlinear geometric structure inherent in data is provided.
However, these operations are typically computationally demanding.
In particular, we focus on Bayesian quadrature (BQ) to numerically compute integrals over normal laws.
We show that by leveraging both prior knowledge and an active exploration scheme, BQ significantly reduces the number of required evaluations.
arXiv Detail & Related papers (2021-02-12T17:38:04Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z) - Gaussianization Flows [113.79542218282282]
We propose a new type of normalizing flow model that enables both efficient iteration of likelihoods and efficient inversion for sample generation.
Because of this guaranteed expressivity, they can capture multimodal target distributions without compromising the efficiency of sample generation.
arXiv Detail & Related papers (2020-03-04T08:15:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.