A Robust Spectral Algorithm for Overcomplete Tensor Decomposition
- URL: http://arxiv.org/abs/2203.02790v1
- Date: Sat, 5 Mar 2022 17:25:37 GMT
- Title: A Robust Spectral Algorithm for Overcomplete Tensor Decomposition
- Authors: Samuel B. Hopkins, Tselil Schramm, Jonathan Shi
- Abstract summary: We give a spectral algorithm for decomposing overcomplete order-4 tensors.
Our algorithm is robust to adversarial perturbations of bounded spectral norm.
Our algorithm avoids semidefinite programming and may be implemented as a series of basic linear-algebraic operations.
- Score: 10.742675209112623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We give a spectral algorithm for decomposing overcomplete order-4 tensors, so
long as their components satisfy an algebraic non-degeneracy condition that
holds for nearly all (all but an algebraic set of measure $0$) tensors over
$(\mathbb{R}^d)^{\otimes 4}$ with rank $n \le d^2$. Our algorithm is robust to
adversarial perturbations of bounded spectral norm.
Our algorithm is inspired by one which uses the sum-of-squares semidefinite
programming hierarchy (Ma, Shi, and Steurer STOC'16, arXiv:1610.01980), and we
achieve comparable robustness and overcompleteness guarantees under similar
algebraic assumptions. However, our algorithm avoids semidefinite programming
and may be implemented as a series of basic linear-algebraic operations. We
consequently obtain a much faster running time than semidefinite programming
methods: our algorithm runs in time $\tilde O(n^2d^3) \le \tilde O(d^7)$, which
is subquadratic in the input size $d^4$ (where we have suppressed factors
related to the condition number of the input tensor).
Related papers
- A Sub-Quadratic Time Algorithm for Robust Sparse Mean Estimation [6.853165736531941]
We study the algorithmic problem of sparse mean estimation in the presence of adversarial outliers.
Our main contribution is an algorithm for robust sparse mean estimation which runs in emphsubquadratic time using $mathrmpoly(k,log d,1/epsilon)$ samples.
arXiv Detail & Related papers (2024-03-07T18:23:51Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Near-Optimal Bounds for Learning Gaussian Halfspaces with Random
Classification Noise [50.64137465792738]
We show that any efficient SQ algorithm for the problem requires sample complexity at least $Omega(d1/2/(maxp, epsilon)2)$.
Our lower bound suggests that this quadratic dependence on $1/epsilon$ is inherent for efficient algorithms.
arXiv Detail & Related papers (2023-07-13T18:59:28Z) - Fast and Practical Quantum-Inspired Classical Algorithms for Solving
Linear Systems [11.929584800629673]
We propose fast and practical quantum-inspired classical algorithms for solving linear systems.
Our main contribution is the application of the heavy ball momentum method to quantum-inspired classical algorithms for solving linear systems.
arXiv Detail & Related papers (2023-07-13T08:46:19Z) - Private estimation algorithms for stochastic block models and mixture
models [63.07482515700984]
General tools for designing efficient private estimation algorithms.
First efficient $(epsilon, delta)$-differentially private algorithm for both weak recovery and exact recovery.
arXiv Detail & Related papers (2023-01-11T09:12:28Z) - Average-Case Complexity of Tensor Decomposition for Low-Degree
Polynomials [93.59919600451487]
"Statistical-computational gaps" occur in many statistical inference tasks.
We consider a model for random order-3 decomposition where one component is slightly larger in norm than the rest.
We show that tensor entries can accurately estimate the largest component when $ll n3/2$ but fail to do so when $rgg n3/2$.
arXiv Detail & Related papers (2022-11-10T00:40:37Z) - Learning a Single Neuron with Adversarial Label Noise via Gradient
Descent [50.659479930171585]
We study a function of the form $mathbfxmapstosigma(mathbfwcdotmathbfx)$ for monotone activations.
The goal of the learner is to output a hypothesis vector $mathbfw$ that $F(mathbbw)=C, epsilon$ with high probability.
arXiv Detail & Related papers (2022-06-17T17:55:43Z) - Sketching Algorithms and Lower Bounds for Ridge Regression [65.0720777731368]
We give a sketching-based iterative algorithm that computes $1+varepsilon$ approximate solutions for the ridge regression problem.
We also show that this algorithm can be used to give faster algorithms for kernel ridge regression.
arXiv Detail & Related papers (2022-04-13T22:18:47Z) - Reconstruction Algorithms for Low-Rank Tensors and Depth-3 Multilinear
Circuits [4.129484350382891]
We give new and efficient black-box reconstruction algorithms for some classes of depth$3$ arithmetic circuits.
Our algorithm works over all fields characteristic 0 or large enough characteristic.
arXiv Detail & Related papers (2021-05-04T20:45:07Z) - Fast Tensor Disentangling Algorithm [0.0]
We introduce an approximate, fast, and simple algorithm to optimize disentangling unitary tensors.
Our algorithm is faster than previous iterative algorithms and often results in a residual entanglement entropy that is within 10 to 40% of the minimum.
arXiv Detail & Related papers (2021-04-16T18:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.