Permanent of random matrices from representation theory: moments,
numerics, concentration, and comments on hardness of boson-sampling
- URL: http://arxiv.org/abs/2104.06423v1
- Date: Tue, 13 Apr 2021 18:00:06 GMT
- Title: Permanent of random matrices from representation theory: moments,
numerics, concentration, and comments on hardness of boson-sampling
- Authors: Sepehr Nezami
- Abstract summary: We study permanents of random i.i.d. complex Gaussian matrices and submatrices of random unitary matrices.
Using a hybrid-theoretic and approach, we prove strong lower bounds for all moments of the permanent distribution.
We show that our bounds are close to being tight and constitute accurate estimates for the moments.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Computing the distribution of permanents of random matrices has been an
outstanding open problem for several decades. In quantum computing,
"anti-concentration" of this distribution is an unproven input for the proof of
hardness of the task of boson-sampling. We study permanents of random i.i.d.
complex Gaussian matrices, and more broadly, submatrices of random unitary
matrices. Using a hybrid representation-theoretic and combinatorial approach,
we prove strong lower bounds for all moments of the permanent distribution. We
provide substantial evidence that our bounds are close to being tight and
constitute accurate estimates for the moments. Let $U(d)^{k\times k}$ be the
distribution of $k\times k$ submatrices of $d\times d$ random unitary matrices,
and $G^{k\times k}$ be the distribution of $k\times k$ complex Gaussian
matrices. (1) Using the Schur-Weyl duality (or the Howe duality), we prove an
expansion formula for the $2t$-th moment of $|Perm(M)|$ when $M$ is drawn from
$U(d)^{k\times k}$ or $G^{k\times k}$. (2) We prove a surprising size-moment
duality: the $2t$-th moment of the permanent of random $k\times k$ matrices is
equal to the $2k$-th moment of the permanent of $t\times t$ matrices. (3) We
design an algorithm to exactly compute high moments of the permanent of small
matrices. (4) We prove lower bounds for arbitrary moments of permanents of
matrices drawn from $G^{ k\times k}$ or $U(k)$, and conjecture that our lower
bounds are close to saturation up to a small multiplicative error. (5) Assuming
our conjectures, we use the large deviation theory to compute the tail of the
distribution of log-permanent of Gaussian matrices for the first time. (6) We
argue that it is unlikely that the permanent distribution can be uniquely
determined from the integer moments and one may need to supplement the moment
calculations with extra assumptions to prove the anti-concentration conjecture.
Related papers
- Private Low-Rank Approximation for Covariance Matrices, Dyson Brownian Motion, and Eigenvalue-Gap Bounds for Gaussian Perturbations [29.212403229351253]
We analyze a complex variant of the Gaussian mechanism and obtain upper bounds on the Frobenius norm of the difference between the matrix output by this mechanism and the best rank-$k$ approximation to $M$.
We show that the eigenvalues of the matrix $M$ perturbed by Gaussian noise have large gaps with high probability.
arXiv Detail & Related papers (2025-02-11T15:46:03Z) - Optimal Quantization for Matrix Multiplication [35.007966885532724]
We build a universal quantizer based on nested lattices with an explicit guarantee of approximation error.
A practical low-complexity version of our quantizer achieves performance quite close to optimal.
arXiv Detail & Related papers (2024-10-17T17:19:48Z) - Efficient Unitary T-designs from Random Sums [0.6640968473398456]
Unitary $T$-designs play an important role in quantum information, with diverse applications in quantum algorithms, benchmarking, tomography, and communication.
We provide a new construction of $T$-designs via random matrix theory using $tildeO(T2 n2)$ quantum gates.
arXiv Detail & Related papers (2024-02-14T17:32:30Z) - Provably learning a multi-head attention layer [55.2904547651831]
Multi-head attention layer is one of the key components of the transformer architecture that sets it apart from traditional feed-forward models.
In this work, we initiate the study of provably learning a multi-head attention layer from random examples.
We prove computational lower bounds showing that in the worst case, exponential dependence on $m$ is unavoidable.
arXiv Detail & Related papers (2024-02-06T15:39:09Z) - Private Covariance Approximation and Eigenvalue-Gap Bounds for Complex
Gaussian Perturbations [28.431572772564518]
We show that the Frobenius norm of the difference between the matrix output by this mechanism and the best rank-$k$ approximation to $M$ is bounded by roughly $tildeO(sqrtkd)$.
This improves on previous work that requires that the gap between every pair of top-$k$ eigenvalues of $M$ is at least $sqrtd$ for a similar bound.
arXiv Detail & Related papers (2023-06-29T03:18:53Z) - A Nearly Tight Bound for Fitting an Ellipsoid to Gaussian Random Points [50.90125395570797]
This nearly establishes a conjecture ofciteSaundersonCPW12, within logarithmic factors.
The latter conjecture has attracted significant attention over the past decade, due to its connections to machine learning and sum-of-squares lower bounds for certain statistical problems.
arXiv Detail & Related papers (2022-12-21T17:48:01Z) - Re-Analyze Gauss: Bounds for Private Matrix Approximation via Dyson
Brownian Motion [28.431572772564518]
Given a symmetric matrix $M$ and a vector $lambda$, we present new bounds on the Frobenius-distance utility of the Gaussian mechanism for approximating $M$ by a matrix.
Our bounds depend on both $lambda$ and the gaps in the eigenvalues of $M$, and hold whenever the top $k+1$ eigenvalues of $M$ have sufficiently large gaps.
arXiv Detail & Related papers (2022-11-11T18:54:01Z) - An Equivalence Principle for the Spectrum of Random Inner-Product Kernel
Matrices with Polynomial Scalings [21.727073594338297]
This study is motivated by applications in machine learning and statistics.
We establish the weak limit of the empirical distribution of these random matrices in a scaling regime.
Our results can be characterized as the free additive convolution between a Marchenko-Pastur law and a semicircle law.
arXiv Detail & Related papers (2022-05-12T18:50:21Z) - Random matrices in service of ML footprint: ternary random features with
no performance loss [55.30329197651178]
We show that the eigenspectrum of $bf K$ is independent of the distribution of the i.i.d. entries of $bf w$.
We propose a novel random technique, called Ternary Random Feature (TRF)
The computation of the proposed random features requires no multiplication and a factor of $b$ less bits for storage compared to classical random features.
arXiv Detail & Related papers (2021-10-05T09:33:49Z) - Spectral properties of sample covariance matrices arising from random
matrices with independent non identically distributed columns [50.053491972003656]
It was previously shown that the functionals $texttr(AR(z))$, for $R(z) = (frac1nXXT- zI_p)-1$ and $Ain mathcal M_p$ deterministic, have a standard deviation of order $O(|A|_* / sqrt n)$.
Here, we show that $|mathbb E[R(z)] - tilde R(z)|_F
arXiv Detail & Related papers (2021-09-06T14:21:43Z) - Linear-Sample Learning of Low-Rank Distributions [56.59844655107251]
We show that learning $ktimes k$, rank-$r$, matrices to normalized $L_1$ distance requires $Omega(frackrepsilon2)$ samples.
We propose an algorithm that uses $cal O(frackrepsilon2log2fracepsilon)$ samples, a number linear in the high dimension, and nearly linear in the matrices, typically low, rank proofs.
arXiv Detail & Related papers (2020-09-30T19:10:32Z) - Optimal Robust Linear Regression in Nearly Linear Time [97.11565882347772]
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = langle X,w* rangle + epsilon$
We propose estimators for this problem under two settings: (i) $X$ is L4-L2 hypercontractive, $mathbbE [XXtop]$ has bounded condition number and $epsilon$ has bounded variance and (ii) $X$ is sub-Gaussian with identity second moment and $epsilon$ is
arXiv Detail & Related papers (2020-07-16T06:44:44Z) - Uniform Probability Distribution Over All Density Matrices [0.0]
We identify a probability measure $u$ on $mathscrD$ that can be regarded as the uniform distribution over $mathscrD$.
We compute the joint distribution of the eigenvalues of a random density matrix distributed according to this measure.
arXiv Detail & Related papers (2020-03-29T17:45:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.