Clustering Mixtures with Almost Optimal Separation in Polynomial Time
- URL: http://arxiv.org/abs/2112.00706v1
- Date: Wed, 1 Dec 2021 18:34:09 GMT
- Title: Clustering Mixtures with Almost Optimal Separation in Polynomial Time
- Authors: Jerry Li, Allen Liu
- Abstract summary: We consider the problem of clustering mixtures of mean-separated Gaussians in high dimensions.
It is folklore that separation $Delta = Theta (sqrtlog k)$ is both necessary and sufficient to recover a good clustering.
We give an algorithm which takes many samples and time, and which can successfully recover a good clustering.
- Score: 21.509452343745252
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of clustering mixtures of mean-separated Gaussians in
high dimensions. We are given samples from a mixture of $k$ identity covariance
Gaussians, so that the minimum pairwise distance between any two pairs of means
is at least $\Delta$, for some parameter $\Delta > 0$, and the goal is to
recover the ground truth clustering of these samples. It is folklore that
separation $\Delta = \Theta (\sqrt{\log k})$ is both necessary and sufficient
to recover a good clustering, at least information theoretically. However, the
estimators which achieve this guarantee are inefficient. We give the first
algorithm which runs in polynomial time, and which almost matches this
guarantee. More precisely, we give an algorithm which takes polynomially many
samples and time, and which can successfully recover a good clustering, so long
as the separation is $\Delta = \Omega (\log^{1/2 + c} k)$, for any $c > 0$.
Previously, polynomial time algorithms were only known for this problem when
the separation was polynomial in $k$, and all algorithms which could tolerate
$\textsf{poly}( \log k )$ separation required quasipolynomial time. We also
extend our result to mixtures of translations of a distribution which satisfies
the Poincar\'{e} inequality, under additional mild assumptions. Our main
technical tool, which we believe is of independent interest, is a novel way to
implicitly represent and estimate high degree moments of a distribution, which
allows us to extract important information about high-degree moments without
ever writing down the full moment tensors explicitly.
Related papers
- Learning general Gaussian mixtures with efficient score matching [16.06356123715737]
We study the problem of learning mixtures of $k$ Gaussians in $d$ dimensions.
We make no separation assumptions on the underlying mixture components.
We give an algorithm that draws $dmathrmpoly(k/varepsilon)$ samples from the target mixture, runs in sample-polynomial time, and constructs a sampler.
arXiv Detail & Related papers (2024-04-29T17:30:36Z) - Clustering Mixtures of Bounded Covariance Distributions Under Optimal
Separation [44.25945344950543]
We study the clustering problem for mixtures of bounded covariance distributions.
We give the first poly-time algorithm for this clustering task.
Our algorithm is robust to $Omega(alpha)$-fraction of adversarial outliers.
arXiv Detail & Related papers (2023-12-19T01:01:53Z) - Do you know what q-means? [50.045011844765185]
Clustering is one of the most important tools for analysis of large datasets.
We present an improved version of the "$q$-means" algorithm for clustering.
We also present a "dequantized" algorithm for $varepsilon which runs in $Obig(frack2varepsilon2(sqrtkd + log(Nd))big.
arXiv Detail & Related papers (2023-08-18T17:52:12Z) - Private estimation algorithms for stochastic block models and mixture
models [63.07482515700984]
General tools for designing efficient private estimation algorithms.
First efficient $(epsilon, delta)$-differentially private algorithm for both weak recovery and exact recovery.
arXiv Detail & Related papers (2023-01-11T09:12:28Z) - Beyond Parallel Pancakes: Quasi-Polynomial Time Guarantees for
Non-Spherical Gaussian Mixtures [9.670578317106182]
We consider mixtures of $kgeq 2$ Gaussian components with unknown means and unknown covariance (identical for all components) that are well-separated.
We show that this kind of hardness can only appear if mixing weights are allowed to be exponentially small.
We develop an algorithm based on the sum-of-squares method with running time quasi-polynomial in the minimum mixing weight.
arXiv Detail & Related papers (2021-12-10T10:51:44Z) - Fuzzy Clustering with Similarity Queries [56.96625809888241]
The fuzzy or soft objective is a popular generalization of the well-known $k$-means problem.
We show that by making few queries, the problem becomes easier to solve.
arXiv Detail & Related papers (2021-06-04T02:32:26Z) - A Polynomial Time Algorithm for Learning Halfspaces with Tsybakov Noise [55.45544638410858]
We study the problem of PAC learning homogeneous halfspaces in the presence of Tsybakov noise.
Our algorithm learns the true halfspace within any accuracy $epsilon$.
arXiv Detail & Related papers (2020-10-04T22:19:06Z) - Computationally efficient sparse clustering [67.95910835079825]
We provide a finite sample analysis of a new clustering algorithm based on PCA.
We show that it achieves the minimax optimal misclustering rate in the regime $|theta infty$.
arXiv Detail & Related papers (2020-05-21T17:51:30Z) - Robustly Learning any Clusterable Mixture of Gaussians [55.41573600814391]
We study the efficient learnability of high-dimensional Gaussian mixtures in the adversarial-robust setting.
We provide an algorithm that learns the components of an $epsilon$-corrupted $k$-mixture within information theoretically near-optimal error proofs of $tildeO(epsilon)$.
Our main technical contribution is a new robust identifiability proof clusters from a Gaussian mixture, which can be captured by the constant-degree Sum of Squares proof system.
arXiv Detail & Related papers (2020-05-13T16:44:12Z) - Outlier-Robust Clustering of Non-Spherical Mixtures [5.863264019032882]
We give the first outlier-robust efficient algorithm for clustering a mixture of $k$ statistically separated d-dimensional Gaussians (k-GMMs)
Our results extend to clustering mixtures of arbitrary affine transforms of the uniform distribution on the $d$-dimensional unit sphere.
arXiv Detail & Related papers (2020-05-06T17:24:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.