Dimension-free Remez Inequalities and norm designs
- URL: http://arxiv.org/abs/2310.07926v5
- Date: Wed, 20 Dec 2023 18:35:23 GMT
- Title: Dimension-free Remez Inequalities and norm designs
- Authors: Lars Becker, Ohad Klein, Joseph Slote, Alexander Volberg, Haonan Zhang
- Abstract summary: A class of domains $X$ and test sets $Y$ -- termed emphnorm -- enjoy dimension-free Remez-type estimates.
We show that the supremum of $f$ does not increase by more than $mathcalO(log K)2d$ when $f$ is extended to the polytorus.
- Score: 48.5897526636987
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The classical Remez inequality bounds the supremum of a bounded-degree
polynomial on an interval $X$ by its supremum on any subset $Y\subset X$ of
positive Lebesgue measure.
There are many multivariate generalizations of the Remez inequality, but most
have constants that depend strongly on dimension.
Here we show that a broad class of domains $X$ and test sets $Y$ -- termed
\emph{norm designs} -- enjoy dimension-free Remez-type estimates.
Instantiations of this theorem allow us for example \emph{a}) to bound the
supremum of an $n$-variate degree-$d$ polynomial on the solid cube $[0,1]^n$ by
its supremum on the regular grid $\{0,1/d,2/d,\ldots, 1\}^n$ independent of
dimension; and \emph{b}) in the case of a degree-$d$ polynomial
$f:\mathbf{Z}_K^n\to\mathbf{C}$ on the $n$-fold product of cyclic groups of
order $K$, to show the supremum of $f$ does not increase by more than
$\mathcal{O}(\log K)^{2d}$ when $f$ is extended to the polytorus as
$f:\mathbf{T}^n\to\mathbf{C}$.
Related papers
- The Communication Complexity of Approximating Matrix Rank [50.6867896228563]
We show that this problem has randomized communication complexity $Omega(frac1kcdot n2log|mathbbF|)$.
As an application, we obtain an $Omega(frac1kcdot n2log|mathbbF|)$ space lower bound for any streaming algorithm with $k$ passes.
arXiv Detail & Related papers (2024-10-26T06:21:42Z) - Provably learning a multi-head attention layer [55.2904547651831]
Multi-head attention layer is one of the key components of the transformer architecture that sets it apart from traditional feed-forward models.
In this work, we initiate the study of provably learning a multi-head attention layer from random examples.
We prove computational lower bounds showing that in the worst case, exponential dependence on $m$ is unavoidable.
arXiv Detail & Related papers (2024-02-06T15:39:09Z) - Noncompact uniform universal approximation [0.0]
The universal approximation theorem is generalised to uniform convergence on the (noncompact) input space $mathbbRn$.
All continuous functions that vanish at infinity can be uniformly approximated by neural networks.
arXiv Detail & Related papers (2023-08-07T08:54:21Z) - Higher rank antipodality [0.0]
Motivated by general probability theory, we say that the set $X$ in $mathbbRd$ is emphantipodal of rank $k$.
For $k=1$, it coincides with the well-studied notion of (pairwise) antipodality introduced by Klee.
arXiv Detail & Related papers (2023-07-31T17:15:46Z) - Quantum and classical low-degree learning via a dimension-free Remez
inequality [52.12931955662553]
We show a new way to relate functions on the hypergrid to their harmonic extensions over the polytorus.
We show the supremum of a function $f$ over products of the cyclic group $exp(2pi i k/K)_k=1K$.
We extend to new spaces a recent line of work citeEI22, CHP, VZ22 that gave similarly efficient methods for learning low-degrees on hypercubes and observables on qubits.
arXiv Detail & Related papers (2023-01-04T04:15:40Z) - Monogamy of entanglement between cones [68.8204255655161]
We show that monogamy is not only a feature of quantum theory, but that it characterizes the minimal tensor product of general pairs of convex cones.
Our proof makes use of a new characterization of products of simplices up to affine equivalence.
arXiv Detail & Related papers (2022-06-23T16:23:59Z) - Low-degree learning and the metric entropy of polynomials [44.99833362998488]
We prove that any (deterministic or randomized) algorithm which learns $mathscrF_nd$ with $L$-accuracy $varepsilon$ requires at least $Omega(sqrtvarepsilon)2dlog n leq log mathsfM(mathscrF_n,d,|cdot|_L,varepsilon) satisfies the two-sided estimate $$c (1-varepsilon)2dlog
arXiv Detail & Related papers (2022-03-17T23:52:08Z) - Mutually unbiased bases: polynomial optimization and symmetry [1.024113475677323]
A set of $k$ orthonormal bases of $mathbb Cd$ is called mutually unbiased $|langle e,frangle |2 = 1/d$ whenever $e$ and $f$ are basis vectors in distinct bases.
We exploit this symmetry (analytically) to reduce the size of the semidefinite programs making them tractable.
arXiv Detail & Related papers (2021-11-10T14:14:53Z) - Linear Bandits on Uniformly Convex Sets [88.3673525964507]
Linear bandit algorithms yield $tildemathcalO(nsqrtT)$ pseudo-regret bounds on compact convex action sets.
Two types of structural assumptions lead to better pseudo-regret bounds.
arXiv Detail & Related papers (2021-03-10T07:33:03Z) - Some convergent results for Backtracking Gradient Descent method on
Banach spaces [0.0]
bf Theorem. Let $X$ be a Banach space and $f:Xrightarrow mathbbR$ be a $C2$ function.
Let $mathcalC$ be the set of critical points of $f$.
arXiv Detail & Related papers (2020-01-16T12:49:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.