Tackling small eigen-gaps: Fine-grained eigenvector estimation and
inference under heteroscedastic noise
- URL: http://arxiv.org/abs/2001.04620v3
- Date: Tue, 7 Sep 2021 21:23:43 GMT
- Title: Tackling small eigen-gaps: Fine-grained eigenvector estimation and
inference under heteroscedastic noise
- Authors: Chen Cheng, Yuting Wei, Yuxin Chen
- Abstract summary: Two fundamental challenges arise in eigenvector estimation and inference for a low-rank matrix from noisy observations.
We propose estimation and uncertainty quantification procedures for an unknown eigenvector.
We establish optimal procedures to construct confidence intervals for the unknown eigenvalues.
- Score: 28.637772416856194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper aims to address two fundamental challenges arising in eigenvector
estimation and inference for a low-rank matrix from noisy observations: (1) how
to estimate an unknown eigenvector when the eigen-gap (i.e. the spacing between
the associated eigenvalue and the rest of the spectrum) is particularly small;
(2) how to perform estimation and inference on linear functionals of an
eigenvector -- a sort of "fine-grained" statistical reasoning that goes far
beyond the usual $\ell_2$ analysis. We investigate how to address these
challenges in a setting where the unknown $n\times n$ matrix is symmetric and
the additive noise matrix contains independent (and non-symmetric) entries.
Based on eigen-decomposition of the asymmetric data matrix, we propose
estimation and uncertainty quantification procedures for an unknown
eigenvector, which further allow us to reason about linear functionals of an
unknown eigenvector. The proposed procedures and the accompanying theory enjoy
several important features: (1) distribution-free (i.e. prior knowledge about
the noise distributions is not needed); (2) adaptive to heteroscedastic noise;
(3) minimax optimal under Gaussian noise. Along the way, we establish optimal
procedures to construct confidence intervals for the unknown eigenvalues. All
this is guaranteed even in the presence of a small eigen-gap (up to
$O(\sqrt{n/\mathrm{poly}\log (n)})$ times smaller than the requirement in prior
theory), which goes significantly beyond what generic matrix perturbation
theory has to offer.
Related papers
- Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Private Covariance Approximation and Eigenvalue-Gap Bounds for Complex
Gaussian Perturbations [28.431572772564518]
We show that the Frobenius norm of the difference between the matrix output by this mechanism and the best rank-$k$ approximation to $M$ is bounded by roughly $tildeO(sqrtkd)$.
This improves on previous work that requires that the gap between every pair of top-$k$ eigenvalues of $M$ is at least $sqrtd$ for a similar bound.
arXiv Detail & Related papers (2023-06-29T03:18:53Z) - General Gaussian Noise Mechanisms and Their Optimality for Unbiased Mean
Estimation [58.03500081540042]
A classical approach to private mean estimation is to compute the true mean and add unbiased, but possibly correlated, Gaussian noise to it.
We show that for every input dataset, an unbiased mean estimator satisfying concentrated differential privacy introduces approximately at least as much error.
arXiv Detail & Related papers (2023-01-31T18:47:42Z) - On confidence intervals for precision matrices and the
eigendecomposition of covariance matrices [20.20416580970697]
This paper tackles the challenge of computing confidence bounds on the individual entries of eigenvectors of a covariance matrix of fixed dimension.
We derive a method to bound the entries of the inverse covariance matrix, the so-called precision matrix.
As an application of these results, we demonstrate a new statistical test, which allows us to test for non-zero values of the precision matrix.
arXiv Detail & Related papers (2022-08-25T10:12:53Z) - Learning Linear Symmetries in Data Using Moment Matching [0.0]
We consider the unsupervised and semi-supervised problems of learning such symmetries in a distribution directly from data.
We show that in the worst case this problem is as difficult as the graph automorphism problem.
We develop and compare theoretically and empirically the effectiveness of different methods of selecting which eigenvectors should have eigenvalue -1 in the symmetry transformation.
arXiv Detail & Related papers (2022-04-04T02:47:37Z) - When Random Tensors meet Random Matrices [50.568841545067144]
This paper studies asymmetric order-$d$ spiked tensor models with Gaussian noise.
We show that the analysis of the considered model boils down to the analysis of an equivalent spiked symmetric textitblock-wise random matrix.
arXiv Detail & Related papers (2021-12-23T04:05:01Z) - Minimax Estimation of Linear Functions of Eigenvectors in the Face of
Small Eigen-Gaps [95.62172085878132]
Eigenvector perturbation analysis plays a vital role in various statistical data science applications.
We develop a suite of statistical theory that characterizes the perturbation of arbitrary linear functions of an unknown eigenvector.
In order to mitigate a non-negligible bias issue inherent to the natural "plug-in" estimator, we develop de-biased estimators.
arXiv Detail & Related papers (2021-04-07T17:55:10Z) - Leveraged Matrix Completion with Noise [84.20092979053119]
We show that we can provably recover an unknown $ntimes n$ matrix of rank $r$ from just about $mathcalO(nrlog2 (n))$ entries.
Our proofs are supported by a novel approach that phrases sufficient optimality conditions based on the Golfing Scheme.
arXiv Detail & Related papers (2020-11-11T16:25:45Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.