Effective Learning of a GMRF Mixture Model
- URL: http://arxiv.org/abs/2005.09030v3
- Date: Fri, 21 Jan 2022 21:32:34 GMT
- Title: Effective Learning of a GMRF Mixture Model
- Authors: Shahaf E. Finder, Eran Treister, Oren Freifeld
- Abstract summary: We propose restricting the GMM to a Gaussian Markov Random Field Mixture Model (GMRF-MM)
When the sparsity pattern of each matrix is known, we propose an efficient optimization method for the Maximum Likelihood Estimate (MLE) of that matrix.
We show that our "debiasing" approach outperforms GLASSO in both the single-GMRF and the GMRF-MM cases.
- Score: 8.336315962271396
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning a Gaussian Mixture Model (GMM) is hard when the number of parameters
is too large given the amount of available data. As a remedy, we propose
restricting the GMM to a Gaussian Markov Random Field Mixture Model (GMRF-MM),
as well as a new method for estimating the latter's sparse precision (i.e.,
inverse covariance) matrices. When the sparsity pattern of each matrix is
known, we propose an efficient optimization method for the Maximum Likelihood
Estimate (MLE) of that matrix. When it is unknown, we utilize the popular
Graphical Least Absolute Shrinkage and Selection Operator (GLASSO) to estimate
that pattern. However, we show that even for a single Gaussian, when GLASSO is
tuned to successfully estimate the sparsity pattern, it does so at the price of
a substantial bias of the values of the nonzero entries of the matrix, and we
show that this problem only worsens in a mixture setting. To overcome this, we
discard the nonzero values estimated by GLASSO, keep only its pattern estimate
and use it within the proposed MLE method. This yields an effective two-step
procedure that removes the bias. We show that our "debiasing" approach
outperforms GLASSO in both the single-GMRF and the GMRF-MM cases. We also show
that when learning priors for image patches, our method outperforms GLASSO even
if we merely use an educated guess about the sparsity pattern, and that our
GMRF-MM outperforms the baseline GMM on real and synthetic high-dimensional
datasets.
Related papers
- Scaling Sparse Fine-Tuning to Large Language Models [67.59697720719672]
Large Language Models (LLMs) are difficult to fully fine-tune due to their sheer number of parameters.
We propose SpIEL, a novel sparse finetuning method which maintains an array of parameter indices and the deltas of these parameters relative to their pretrained values.
We show that SpIEL is superior to popular parameter-efficient fine-tuning methods like LoRA in terms of performance and comparable in terms of run time.
arXiv Detail & Related papers (2024-01-29T18:43:49Z) - Algorithme EM r\'egularis\'e [0.0]
This paper presents a regularized version of the EM algorithm that efficiently uses prior knowledge to cope with a small sample size.
Experiments on real data highlight the good performance of the proposed algorithm for clustering purposes.
arXiv Detail & Related papers (2023-07-04T23:19:25Z) - Regularized EM algorithm [9.367612782346205]
We present a regularized EM algorithm for GMM-s that can make efficient use of such prior knowledge as well as cope with LSS situations.
We show that the theoretical guarantees of convergence hold, leading to better performing EM algorithm for structured covariance matrix models or with low sample settings.
arXiv Detail & Related papers (2023-03-27T08:32:20Z) - A distribution-free mixed-integer optimization approach to hierarchical modelling of clustered and longitudinal data [0.0]
We introduce an innovative algorithm that evaluates cluster effects for new data points, thereby increasing the robustness and precision of this model.
The inferential and predictive efficacy of this approach is further illustrated through its application in student scoring and protein expression.
arXiv Detail & Related papers (2023-02-06T23:34:51Z) - Stochastic First-Order Learning for Large-Scale Flexibly Tied Gaussian
Mixture Model [3.4546761246181696]
We propose a new optimization algorithm on the manifold of Gaussian Mixture Models (GMMs)
We observe that methods can outperform the expectation-maximization algorithm in terms of attaining better likelihood, needing fewer epochs for convergence, and consuming less time per each epoch.
arXiv Detail & Related papers (2022-12-11T04:24:52Z) - Langevin Monte Carlo for Contextual Bandits [72.00524614312002]
Langevin Monte Carlo Thompson Sampling (LMC-TS) is proposed to directly sample from the posterior distribution in contextual bandits.
We prove that the proposed algorithm achieves the same sublinear regret bound as the best Thompson sampling algorithms for a special case of contextual bandits.
arXiv Detail & Related papers (2022-06-22T17:58:23Z) - Continual Learning with Fully Probabilistic Models [70.3497683558609]
We present an approach for continual learning based on fully probabilistic (or generative) models of machine learning.
We propose a pseudo-rehearsal approach using a Gaussian Mixture Model (GMM) instance for both generator and classifier functionalities.
We show that GMR achieves state-of-the-art performance on common class-incremental learning problems at very competitive time and memory complexity.
arXiv Detail & Related papers (2021-04-19T12:26:26Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Robust Compressed Sensing using Generative Models [98.64228459705859]
In this paper we propose an algorithm inspired by the Median-of-Means (MOM)
Our algorithm guarantees recovery for heavy-tailed data, even in the presence of outliers.
arXiv Detail & Related papers (2020-06-16T19:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.