Robust $k$-means Clustering for Distributions with Two Moments
- URL: http://arxiv.org/abs/2002.02339v2
- Date: Tue, 3 Nov 2020 15:59:41 GMT
- Title: Robust $k$-means Clustering for Distributions with Two Moments
- Authors: Yegor Klochkov, Alexey Kroshnin, and Nikita Zhivotovskiy
- Abstract summary: We consider the robust algorithms for the $k$-means clustering problem where a quantizer is constructed based on $N$ independent observations.
Our main results are median of means based non-asymptotic excess distortion bounds that hold under the two bounded moments assumption in a general separable Hilbert space.
- Score: 4.21934751979057
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the robust algorithms for the $k$-means clustering problem where
a quantizer is constructed based on $N$ independent observations. Our main
results are median of means based non-asymptotic excess distortion bounds that
hold under the two bounded moments assumption in a general separable Hilbert
space. In particular, our results extend the renowned asymptotic result of
Pollard, 1981 who showed that the existence of two moments is sufficient for
strong consistency of an empirically optimal quantizer in $\mathbb{R}^d$. In a
special case of clustering in $\mathbb{R}^d$, under two bounded moments, we
prove matching (up to constant factors) non-asymptotic upper and lower bounds
on the excess distortion, which depend on the probability mass of the lightest
cluster of an optimal quantizer. Our bounds have the sub-Gaussian form, and the
proofs are based on the versions of uniform bounds for robust mean estimators.
Related papers
- Breaking the Heavy-Tailed Noise Barrier in Stochastic Optimization Problems [56.86067111855056]
We consider clipped optimization problems with heavy-tailed noise with structured density.
We show that it is possible to get faster rates of convergence than $mathcalO(K-(alpha - 1)/alpha)$, when the gradients have finite moments of order.
We prove that the resulting estimates have negligible bias and controllable variance.
arXiv Detail & Related papers (2023-11-07T17:39:17Z) - Lattice-Based Methods Surpass Sum-of-Squares in Clustering [98.46302040220395]
Clustering is a fundamental primitive in unsupervised learning.
Recent work has established lower bounds against the class of low-degree methods.
We show that, perhaps surprisingly, this particular clustering model textitdoes not exhibit a statistical-to-computational gap.
arXiv Detail & Related papers (2021-12-07T18:50:17Z) - Uncertainty quantification in the Bradley-Terry-Luce model [14.994932962403935]
This paper focuses on two estimators that have received much recent attention: the maximum likelihood estimator (MLE) and the spectral estimator.
Using a unified proof strategy, we derive sharp and uniform non-asymptotic expansions for both estimators in the sparsest possible regime.
Our proof is based on a self-consistent equation of the second-order vector and a novel leave-two-out analysis.
arXiv Detail & Related papers (2021-10-08T03:06:30Z) - Optimal policy evaluation using kernel-based temporal difference methods [78.83926562536791]
We use kernel Hilbert spaces for estimating the value function of an infinite-horizon discounted Markov reward process.
We derive a non-asymptotic upper bound on the error with explicit dependence on the eigenvalues of the associated kernel operator.
We prove minimax lower bounds over sub-classes of MRPs.
arXiv Detail & Related papers (2021-09-24T14:48:20Z) - Spatially relaxed inference on high-dimensional linear models [48.989769153211995]
We study the properties of ensembled clustered inference algorithms which combine spatially constrained clustering, statistical inference, and ensembling to aggregate several clustered inference solutions.
We show that ensembled clustered inference algorithms control the $delta$-FWER under standard assumptions for $delta$ equal to the largest cluster diameter.
arXiv Detail & Related papers (2021-06-04T16:37:19Z) - Exact Recovery in the General Hypergraph Stochastic Block Model [92.28929858529679]
This paper investigates fundamental limits of exact recovery in the general d-uniform hypergraph block model (d-HSBM)
We show that there exists a sharp threshold such that exact recovery is achievable above the threshold and impossible below it.
arXiv Detail & Related papers (2021-05-11T03:39:08Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - Convergence Rates of Stochastic Gradient Descent under Infinite Noise
Variance [14.06947898164194]
Heavy tails emerge in gradient descent (SGD) in various scenarios.
We provide convergence guarantees for SGD under a state-dependent and heavy-tailed noise with a potentially infinite variance.
Our results indicate that even under heavy-tailed noise with infinite variance, SGD can converge to the global optimum.
arXiv Detail & Related papers (2021-02-20T13:45:11Z) - Nonparametric approximation of conditional expectation operators [0.3655021726150368]
We investigate the approximation of the $L2$-operator defined by $[Pf](x) := mathbbE[ f(Y) mid X = x ]$ under minimal assumptions.
We prove that $P$ can be arbitrarily well approximated in operator norm by Hilbert-Schmidt operators acting on a reproducing kernel space.
arXiv Detail & Related papers (2020-12-23T19:06:12Z) - Optimal Bounds between $f$-Divergences and Integral Probability Metrics [8.401473551081748]
Families of $f$-divergences and Integral Probability Metrics are widely used to quantify similarity between probability distributions.
We systematically study the relationship between these two families from the perspective of convex duality.
We obtain new bounds while also recovering in a unified manner well-known results, such as Hoeffding's lemma.
arXiv Detail & Related papers (2020-06-10T17:39:11Z) - The role of regularization in classification of high-dimensional noisy
Gaussian mixture [36.279288150100875]
We consider a high-dimensional mixture of two Gaussians in the noisy regime.
We provide a rigorous analysis of the generalization error of regularized convex classifiers.
We discuss surprising effects of the regularization that in some cases allows to reach the Bayes-optimal performances.
arXiv Detail & Related papers (2020-02-26T14:54:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.