Minimax Estimation of Linear Functions of Eigenvectors in the Face of
Small Eigen-Gaps
- URL: http://arxiv.org/abs/2104.03298v1
- Date: Wed, 7 Apr 2021 17:55:10 GMT
- Title: Minimax Estimation of Linear Functions of Eigenvectors in the Face of
Small Eigen-Gaps
- Authors: Gen Li, Changxiao Cai, Yuantao Gu, H. Vincent Poor, Yuxin Chen
- Abstract summary: Eigenvector perturbation analysis plays a vital role in various statistical data science applications.
We develop a suite of statistical theory that characterizes the perturbation of arbitrary linear functions of an unknown eigenvector.
In order to mitigate a non-negligible bias issue inherent to the natural "plug-in" estimator, we develop de-biased estimators.
- Score: 95.62172085878132
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Eigenvector perturbation analysis plays a vital role in various statistical
data science applications. A large body of prior works, however, focused on
establishing $\ell_{2}$ eigenvector perturbation bounds, which are often highly
inadequate in addressing tasks that rely on fine-grained behavior of an
eigenvector. This paper makes progress on this by studying the perturbation of
linear functions of an unknown eigenvector. Focusing on two fundamental
problems -- matrix denoising and principal component analysis -- in the
presence of Gaussian noise, we develop a suite of statistical theory that
characterizes the perturbation of arbitrary linear functions of an unknown
eigenvector. In order to mitigate a non-negligible bias issue inherent to the
natural "plug-in" estimator, we develop de-biased estimators that (1) achieve
minimax lower bounds for a family of scenarios (modulo some logarithmic
factor), and (2) can be computed in a data-driven manner without sample
splitting. Noteworthily, the proposed estimators are nearly minimax optimal
even when the associated eigen-gap is substantially smaller than what is
required in prior theory.
Related papers
- Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Robust Regularized Locality Preserving Indexing for Fiedler Vector
Estimation [32.26669925809068]
In real-world applications, the data may be subject to heavy-tailed noise and outliers which results in deteriorations in the structure of the Fiedler vector estimate.
We design a Robust Regularized Locality Preserving Indexing (RRLPI) method for Fiedler vector estimation that aims to approximate the nonlinear manifold structure of the Laplace Beltrami operator.
arXiv Detail & Related papers (2021-07-26T09:49:23Z) - Fine-grained Generalization Analysis of Vector-valued Learning [28.722350261462463]
We start the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size.
To understand the interaction between optimization and learning, we further use our results to derive the first bounds for descent with vector-valued functions.
As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function.
arXiv Detail & Related papers (2021-04-29T07:57:34Z) - Analysis of Truncated Orthogonal Iteration for Sparse Eigenvector
Problems [78.95866278697777]
We propose two variants of the Truncated Orthogonal Iteration to compute multiple leading eigenvectors with sparsity constraints simultaneously.
We then apply our algorithms to solve the sparse principle component analysis problem for a wide range of test datasets.
arXiv Detail & Related papers (2021-03-24T23:11:32Z) - Probabilistic Simplex Component Analysis [66.30587591100566]
PRISM is a probabilistic simplex component analysis approach to identifying the vertices of a data-circumscribing simplex from data.
The problem has a rich variety of applications, the most notable being hyperspectral unmixing in remote sensing and non-negative matrix factorization in machine learning.
arXiv Detail & Related papers (2021-03-18T05:39:00Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Semiparametric Nonlinear Bipartite Graph Representation Learning with
Provable Guarantees [106.91654068632882]
We consider the bipartite graph and formalize its representation learning problem as a statistical estimation problem of parameters in a semiparametric exponential family distribution.
We show that the proposed objective is strongly convex in a neighborhood around the ground truth, so that a gradient descent-based method achieves linear convergence rate.
Our estimator is robust to any model misspecification within the exponential family, which is validated in extensive experiments.
arXiv Detail & Related papers (2020-03-02T16:40:36Z) - Tackling small eigen-gaps: Fine-grained eigenvector estimation and
inference under heteroscedastic noise [28.637772416856194]
Two fundamental challenges arise in eigenvector estimation and inference for a low-rank matrix from noisy observations.
We propose estimation and uncertainty quantification procedures for an unknown eigenvector.
We establish optimal procedures to construct confidence intervals for the unknown eigenvalues.
arXiv Detail & Related papers (2020-01-14T04:26:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.