All-or-nothing statistical and computational phase transitions in sparse
spiked matrix estimation
- URL: http://arxiv.org/abs/2006.07971v2
- Date: Fri, 30 Oct 2020 11:07:35 GMT
- Title: All-or-nothing statistical and computational phase transitions in sparse
spiked matrix estimation
- Authors: Jean Barbier, Nicolas Macris and Cynthia Rush
- Abstract summary: We analyze the approximate message passing algorithm in a sparse regime.
We find all-or-nothing phase transitions for the minimum and mean-square errors.
In the sparse regime the statistical-to-algorithmic gap diverges indicating that sparse recovery is hard for approximate message passing.
- Score: 35.035853993422506
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We determine statistical and computational limits for estimation of a
rank-one matrix (the spike) corrupted by an additive gaussian noise matrix, in
a sparse limit, where the underlying hidden vector (that constructs the
rank-one matrix) has a number of non-zero components that scales sub-linearly
with the total dimension of the vector, and the signal-to-noise ratio tends to
infinity at an appropriate speed. We prove explicit low-dimensional variational
formulas for the asymptotic mutual information between the spike and the
observed noisy matrix and analyze the approximate message passing algorithm in
the sparse regime. For Bernoulli and Bernoulli-Rademacher distributed vectors,
and when the sparsity and signal strength satisfy an appropriate scaling
relation, we find all-or-nothing phase transitions for the asymptotic minimum
and algorithmic mean-square errors. These jump from their maximum possible
value to zero, at well defined signal-to-noise thresholds whose asymptotic
values we determine exactly. In the asymptotic regime the
statistical-to-algorithmic gap diverges indicating that sparse recovery is hard
for approximate message passing.
Related papers
- Matrix Denoising with Doubly Heteroscedastic Noise: Fundamental Limits and Optimal Spectral Methods [24.06775799553418]
We study the matrix denoising problem of estimating the singular vectors of a rank-$1$ signal corrupted by noise with both column and row correlations.
Our work establishes the information-theoretic and algorithmic limits of matrix denoising with doubly heteroscedastic noise.
arXiv Detail & Related papers (2024-05-22T18:38:10Z) - Provably Accelerating Ill-Conditioned Low-rank Estimation via Scaled
Gradient Descent, Even with Overparameterization [48.65416821017865]
This chapter introduces a new algorithmic approach, dubbed scaled gradient (ScaledGD)
It converges linearly at a constant rate independent of the condition number of the low-rank object.
It maintains the low periteration cost of gradient descent for a variety of tasks.
arXiv Detail & Related papers (2023-10-09T21:16:57Z) - Subspace clustering in high-dimensions: Phase transitions \&
Statistical-to-Computational gap [24.073221004661427]
A simple model to study subspace clustering is the high-dimensional $k$-Gaussian mixture model.
We provide an exact characterization of the statistically optimal reconstruction error in this model in the high-dimensional regime with extensive sparsity.
arXiv Detail & Related papers (2022-05-26T17:47:35Z) - Optimal Estimation and Computational Limit of Low-rank Gaussian Mixtures [12.868722327487752]
We propose a low-rank Gaussian mixture model (LrMM) assuming each matrix-valued observation has a planted low-rank structure.
We prove the minimax optimality of a maximum likelihood estimator which, in general, is computationally infeasible.
Our results reveal multiple phase transitions in the minimax error rates and the statistical-to-computational gap.
arXiv Detail & Related papers (2022-01-22T12:43:25Z) - Nonconvex Stochastic Scaled-Gradient Descent and Generalized Eigenvector
Problems [98.34292831923335]
Motivated by the problem of online correlation analysis, we propose the emphStochastic Scaled-Gradient Descent (SSD) algorithm.
We bring these ideas together in an application to online correlation analysis, deriving for the first time an optimal one-time-scale algorithm with an explicit rate of local convergence to normality.
arXiv Detail & Related papers (2021-12-29T18:46:52Z) - Long Random Matrices and Tensor Unfolding [4.83420384410068]
We prove there exists a critical signal-to-noise ratio depending on the dimensions of the matrix.
As a main application, we investigate the tensor unfolding algorithm for the asymmetric rank-one spiked tensor model.
arXiv Detail & Related papers (2021-10-19T19:13:27Z) - Rank-one matrix estimation with groupwise heteroskedasticity [5.202966939338455]
We study the problem of estimating a rank-one matrix from Gaussian observations where different blocks of the matrix are observed under different noise levels.
We prove exact formulas for the minimum mean-squared error in estimating both the matrix and the latent variables.
We derive an approximate message passing algorithm and a gradient descent algorithm and show empirically that these algorithms achieve the information-theoretic limits in certain regimes.
arXiv Detail & Related papers (2021-06-22T17:48:36Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - ROOT-SGD: Sharp Nonasymptotics and Near-Optimal Asymptotics in a Single Algorithm [71.13558000599839]
We study the problem of solving strongly convex and smooth unconstrained optimization problems using first-order algorithms.
We devise a novel, referred to as Recursive One-Over-T SGD, based on an easily implementable, averaging of past gradients.
We prove that it simultaneously achieves state-of-the-art performance in both a finite-sample, nonasymptotic sense and an sense.
arXiv Detail & Related papers (2020-08-28T14:46:56Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - On Linear Stochastic Approximation: Fine-grained Polyak-Ruppert and
Non-Asymptotic Concentration [115.1954841020189]
We study the inequality and non-asymptotic properties of approximation procedures with Polyak-Ruppert averaging.
We prove a central limit theorem (CLT) for the averaged iterates with fixed step size and number of iterations going to infinity.
arXiv Detail & Related papers (2020-04-09T17:54:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.