Optimal Spectral Recovery of a Planted Vector in a Subspace
- URL: http://arxiv.org/abs/2105.15081v1
- Date: Mon, 31 May 2021 16:10:49 GMT
- Title: Optimal Spectral Recovery of a Planted Vector in a Subspace
- Authors: Cheng Mao, Alexander S. Wein
- Abstract summary: We study efficient estimation and detection of a planted vector $v$ whose $ell_4$ norm differs from that of a Gaussian vector with the same $ell$ norm.
We show that in the regime $n rho gg sqrtN$, any spectral method from a large class (and more generally, any low-degree of the input) fails to detect the planted vector.
- Score: 80.02218763267992
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recovering a planted vector $v$ in an $n$-dimensional random subspace of
$\mathbb{R}^N$ is a generic task related to many problems in machine learning
and statistics, such as dictionary learning, subspace recovery, and principal
component analysis. In this work, we study computationally efficient estimation
and detection of a planted vector $v$ whose $\ell_4$ norm differs from that of
a Gaussian vector with the same $\ell_2$ norm. For instance, in the special
case of an $N \rho$-sparse vector $v$ with Rademacher nonzero entries, our
results include the following:
(1) We give an improved analysis of (a slight variant of) the spectral method
proposed by Hopkins, Schramm, Shi, and Steurer, showing that it approximately
recovers $v$ with high probability in the regime $n \rho \ll \sqrt{N}$. In
contrast, previous work required either $\rho \ll 1/\sqrt{n}$ or $n \sqrt{\rho}
\lesssim \sqrt{N}$ for polynomial-time recovery. Our result subsumes both of
these conditions (up to logarithmic factors) and also treats the dense case
$\rho = 1$ which was not previously considered.
(2) Akin to $\ell_\infty$ bounds for eigenvector perturbation, we establish
an entrywise error bound for the spectral estimator via a leave-one-out
analysis, from which it follows that thresholding recovers $v$ exactly.
(3) We study the associated detection problem and show that in the regime $n
\rho \gg \sqrt{N}$, any spectral method from a large class (and more generally,
any low-degree polynomial of the input) fails to detect the planted vector.
This establishes optimality of our upper bounds and offers evidence that no
polynomial-time algorithm can succeed when $n \rho \gg \sqrt{N}$.
Related papers
- Optimal Sketching for Residual Error Estimation for Matrix and Vector Norms [50.15964512954274]
We study the problem of residual error estimation for matrix and vector norms using a linear sketch.
We demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work.
We also show an $Omega(k2/pn1-2/p)$ lower bound for the sparse recovery problem, which is tight up to a $mathrmpoly(log n)$ factor.
arXiv Detail & Related papers (2024-08-16T02:33:07Z) - A Nearly-Optimal Bound for Fast Regression with $\ell_\infty$ Guarantee [16.409210914237086]
Given a matrix $Ain mathbbRntimes d$ and a tensor $bin mathbbRn$, we consider the regression problem with $ell_infty$ guarantees.
We show that in order to obtain such $ell_infty$ guarantee for $ell$ regression, one has to use sketching matrices that are dense.
We also develop a novel analytical framework for $ell_infty$ guarantee regression that utilizes the Oblivious Coordinate-wise Embedding (OCE) property
arXiv Detail & Related papers (2023-02-01T05:22:40Z) - Overparametrized linear dimensionality reductions: From projection
pursuit to two-layer neural networks [10.368585938419619]
Given a cloud of $n$ data points in $mathbbRd$, consider all projections onto $m$-dimensional subspaces of $mathbbRd$.
What does this collection of probability distributions look like when $n,d$ grow large?
Denoting by $mathscrF_m, alpha$ the set of probability distributions in $mathbbRm$ that arise as low-dimensional projections in this limit, we establish new inner and outer bounds on $mathscrF_
arXiv Detail & Related papers (2022-06-14T00:07:33Z) - Active Sampling for Linear Regression Beyond the $\ell_2$ Norm [70.49273459706546]
We study active sampling algorithms for linear regression, which aim to query only a small number of entries of a target vector.
We show that this dependence on $d$ is optimal, up to logarithmic factors.
We also provide the first total sensitivity upper bound $O(dmax1,p/2log2 n)$ for loss functions with at most degree $p$ growth.
arXiv Detail & Related papers (2021-11-09T00:20:01Z) - Spectral properties of sample covariance matrices arising from random
matrices with independent non identically distributed columns [50.053491972003656]
It was previously shown that the functionals $texttr(AR(z))$, for $R(z) = (frac1nXXT- zI_p)-1$ and $Ain mathcal M_p$ deterministic, have a standard deviation of order $O(|A|_* / sqrt n)$.
Here, we show that $|mathbb E[R(z)] - tilde R(z)|_F
arXiv Detail & Related papers (2021-09-06T14:21:43Z) - Gap-Dependent Unsupervised Exploration for Reinforcement Learning [40.990467706237396]
We present an efficient algorithm for task-agnostic reinforcement learning.
The algorithm takes only $widetildemathcalO (1/epsilon cdot (H3SA / rho + H4 S2 A) )$ episodes of exploration.
We show that, information-theoretically, this bound is nearly tight for $rho Theta (1/(HS))$ and $H>1$.
arXiv Detail & Related papers (2021-08-11T20:42:46Z) - Learning a Latent Simplex in Input-Sparsity Time [58.30321592603066]
We consider the problem of learning a latent $k$-vertex simplex $KsubsetmathbbRdtimes n$, given access to $AinmathbbRdtimes n$.
We show that the dependence on $k$ in the running time is unnecessary given a natural assumption about the mass of the top $k$ singular values of $A$.
arXiv Detail & Related papers (2021-05-17T16:40:48Z) - The Average-Case Time Complexity of Certifying the Restricted Isometry
Property [66.65353643599899]
In compressed sensing, the restricted isometry property (RIP) on $M times N$ sensing matrices guarantees efficient reconstruction of sparse vectors.
We investigate the exact average-case time complexity of certifying the RIP property for $Mtimes N$ matrices with i.i.d. $mathcalN(0,1/M)$ entries.
arXiv Detail & Related papers (2020-05-22T16:55:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.