On Learning Mixture Models with Sparse Parameters
- URL: http://arxiv.org/abs/2202.11940v1
- Date: Thu, 24 Feb 2022 07:44:23 GMT
- Title: On Learning Mixture Models with Sparse Parameters
- Authors: Arya Mazumdar, Soumyabrata Pal
- Abstract summary: We study mixtures with high dimensional sparse latent parameter vectors and consider the problem of support recovery of those vectors.
We provide efficient algorithms for support recovery that have a logarithmic sample complexity dependence on the dimensionality of the latent space.
- Score: 44.3425205248937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mixture models are widely used to fit complex and multimodal datasets. In
this paper we study mixtures with high dimensional sparse latent parameter
vectors and consider the problem of support recovery of those vectors. While
parameter learning in mixture models is well-studied, the sparsity constraint
remains relatively unexplored. Sparsity of parameter vectors is a natural
constraint in variety of settings, and support recovery is a major step towards
parameter estimation. We provide efficient algorithms for support recovery that
have a logarithmic sample complexity dependence on the dimensionality of the
latent space. Our algorithms are quite general, namely they are applicable to
1) mixtures of many different canonical distributions including Uniform,
Poisson, Laplace, Gaussians, etc. 2) Mixtures of linear regressions and linear
classifiers with Gaussian covariates under different assumptions on the unknown
parameters. In most of these settings, our results are the first guarantees on
the problem while in the rest, our results provide improvements on existing
works.
Related papers
- Multivariate root-n-consistent smoothing parameter free matching estimators and estimators of inverse density weighted expectations [51.000851088730684]
We develop novel modifications of nearest-neighbor and matching estimators which converge at the parametric $sqrt n $-rate.
We stress that our estimators do not involve nonparametric function estimators and in particular do not rely on sample-size dependent parameters smoothing.
arXiv Detail & Related papers (2024-07-11T13:28:34Z) - Robust scalable initialization for Bayesian variational inference with
multi-modal Laplace approximations [0.0]
Variational mixtures with full-covariance structures suffer from a quadratic growth due to variational parameters with the number of parameters.
We propose a method for constructing an initial Gaussian model approximation that can be used to warm-start variational inference.
arXiv Detail & Related papers (2023-07-12T19:30:04Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Convex Latent Effect Logit Model via Sparse and Low-rank Decomposition [2.1915057426589746]
We propose a convexparametric convexparametric formulation for learning logistic regression model (logit) with latent heterogeneous effect on sub-population.
Despite its popularity, the mixed logit approach for learning individual heterogeneity has several downsides.
arXiv Detail & Related papers (2021-08-22T22:23:39Z) - Learning Gaussian Mixtures with Generalised Linear Models: Precise
Asymptotics in High-dimensions [79.35722941720734]
Generalised linear models for multi-class classification problems are one of the fundamental building blocks of modern machine learning tasks.
We prove exacts characterising the estimator in high-dimensions via empirical risk minimisation.
We discuss how our theory can be applied beyond the scope of synthetic data.
arXiv Detail & Related papers (2021-06-07T16:53:56Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - Fair Sparse Regression with Clustering: An Invex Relaxation for a
Combinatorial Problem [32.18449686637963]
We show that the inclusion of the debiasing/fairness constraint in our model has no adverse effect on the performance.
We simultaneously solve the clustering problem by recovering the exact value of the hidden attribute for each sample.
arXiv Detail & Related papers (2021-02-19T01:46:34Z) - Bayesian multiscale deep generative model for the solution of
high-dimensional inverse problems [0.0]
A novel multiscale Bayesian inference approach is introduced based on deep probabilistic generative models.
The method allows high-dimensional parameter estimation while exhibiting stability, efficiency and accuracy.
arXiv Detail & Related papers (2021-02-04T11:47:21Z) - Learning Mixtures of Low-Rank Models [89.39877968115833]
We study the problem of learning computational mixtures of low-rank models.
We develop an algorithm that is guaranteed to recover the unknown matrices with near-optimal sample.
In addition, the proposed algorithm is provably stable against random noise.
arXiv Detail & Related papers (2020-09-23T17:53:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.