Beyond Independent Measurements: General Compressed Sensing with GNN
Application
- URL: http://arxiv.org/abs/2111.00327v1
- Date: Sat, 30 Oct 2021 20:35:56 GMT
- Title: Beyond Independent Measurements: General Compressed Sensing with GNN
Application
- Authors: Alireza Naderi and Yaniv Plan
- Abstract summary: We consider the problem of recovering a structured signal $mathbfx in mathbbRn$ from noisy cone observations.
We show that the effective rank of $mathbfB$ may be used as a surrogate for the number of measurements.
- Score: 4.924126492174801
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We consider the problem of recovering a structured signal $\mathbf{x} \in
\mathbb{R}^{n}$ from noisy linear observations $\mathbf{y} =\mathbf{M}
\mathbf{x}+\mathbf{w}$. The measurement matrix is modeled as $\mathbf{M} =
\mathbf{B}\mathbf{A}$, where $\mathbf{B} \in \mathbb{R}^{l \times m}$ is
arbitrary and $\mathbf{A} \in \mathbb{R}^{m \times n}$ has independent
sub-gaussian rows. By varying $\mathbf{B}$, and the sub-gaussian distribution
of $\mathbf{A}$, this gives a family of measurement matrices which may have
heavy tails, dependent rows and columns, and singular values with a large
dynamic range. When the structure is given as a possibly non-convex cone $T
\subset \mathbb{R}^{n}$, an approximate empirical risk minimizer is proven to
be a robust estimator if the effective number of measurements is sufficient,
even in the presence of a model mismatch. In classical compressed sensing with
independent (sub-)gaussian measurements, one asks how many measurements are
needed to recover $\mathbf{x}$? In our setting, however, the effective number
of measurements depends on the properties of $\mathbf{B}$. We show that the
effective rank of $\mathbf{B}$ may be used as a surrogate for the number of
measurements, and if this exceeds the squared Gaussian mean width of $(T-T)
\cap \mathbb{S}^{n-1}$, then accurate recovery is guaranteed. Furthermore, we
examine the special case of generative priors in detail, that is when
$\mathbf{x}$ lies close to $T = \mathrm{ran}(G)$ and $G: \mathbb{R}^k
\rightarrow \mathbb{R}^n$ is a Generative Neural Network (GNN) with ReLU
activation functions. Our work relies on a recent result in random matrix
theory by Jeong, Li, Plan, and Yilmaz arXiv:2001.10631. .
Related papers
- Provably learning a multi-head attention layer [55.2904547651831]
Multi-head attention layer is one of the key components of the transformer architecture that sets it apart from traditional feed-forward models.
In this work, we initiate the study of provably learning a multi-head attention layer from random examples.
We prove computational lower bounds showing that in the worst case, exponential dependence on $m$ is unavoidable.
arXiv Detail & Related papers (2024-02-06T15:39:09Z) - Optimal Estimator for Linear Regression with Shuffled Labels [17.99906229036223]
This paper considers the task of linear regression with shuffled labels.
$mathbf Y in mathbb Rntimes m, mathbf Pi in mathbb Rntimes p, mathbf B in mathbb Rptimes m$, and $mathbf Win mathbb Rntimes m$, respectively.
arXiv Detail & Related papers (2023-10-02T16:44:47Z) - A Unified Framework for Uniform Signal Recovery in Nonlinear Generative
Compressed Sensing [68.80803866919123]
Under nonlinear measurements, most prior results are non-uniform, i.e., they hold with high probability for a fixed $mathbfx*$ rather than for all $mathbfx*$ simultaneously.
Our framework accommodates GCS with 1-bit/uniformly quantized observations and single index models as canonical examples.
We also develop a concentration inequality that produces tighter bounds for product processes whose index sets have low metric entropy.
arXiv Detail & Related papers (2023-09-25T17:54:19Z) - Learning a Single Neuron with Adversarial Label Noise via Gradient
Descent [50.659479930171585]
We study a function of the form $mathbfxmapstosigma(mathbfwcdotmathbfx)$ for monotone activations.
The goal of the learner is to output a hypothesis vector $mathbfw$ that $F(mathbbw)=C, epsilon$ with high probability.
arXiv Detail & Related papers (2022-06-17T17:55:43Z) - Random matrices in service of ML footprint: ternary random features with
no performance loss [55.30329197651178]
We show that the eigenspectrum of $bf K$ is independent of the distribution of the i.i.d. entries of $bf w$.
We propose a novel random technique, called Ternary Random Feature (TRF)
The computation of the proposed random features requires no multiplication and a factor of $b$ less bits for storage compared to classical random features.
arXiv Detail & Related papers (2021-10-05T09:33:49Z) - Spectral properties of sample covariance matrices arising from random
matrices with independent non identically distributed columns [50.053491972003656]
It was previously shown that the functionals $texttr(AR(z))$, for $R(z) = (frac1nXXT- zI_p)-1$ and $Ain mathcal M_p$ deterministic, have a standard deviation of order $O(|A|_* / sqrt n)$.
Here, we show that $|mathbb E[R(z)] - tilde R(z)|_F
arXiv Detail & Related papers (2021-09-06T14:21:43Z) - On the computational and statistical complexity of over-parameterized
matrix sensing [30.785670369640872]
We consider solving the low rank matrix sensing problem with Factorized Gradient Descend (FGD) method.
By decomposing the factorized matrix $mathbfF$ into separate column spaces, we show that $|mathbfF_t - mathbfF_t - mathbfX*|_F2$ converges to a statistical error.
arXiv Detail & Related papers (2021-01-27T04:23:49Z) - Near-Optimal SQ Lower Bounds for Agnostically Learning Halfspaces and
ReLUs under Gaussian Marginals [49.60752558064027]
We study the fundamental problems of agnostically learning halfspaces and ReLUs under Gaussian marginals.
Our lower bounds provide strong evidence that current upper bounds for these tasks are essentially best possible.
arXiv Detail & Related papers (2020-06-29T17:10:10Z) - Efficient Statistics for Sparse Graphical Models from Truncated Samples [19.205541380535397]
We focus on two fundamental and classical problems: (i) inference of sparse Gaussian graphical models and (ii) support recovery of sparse linear models.
For sparse linear regression, suppose samples $(bf x,y)$ are generated where $y = bf xtopOmega* + mathcalN(0,1)$ and $(bf x, y)$ is seen only if $y$ belongs to a truncation set $S subseteq mathbbRd$.
arXiv Detail & Related papers (2020-06-17T09:21:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.