Local Borsuk-Ulam, Stability, and Replicability
- URL: http://arxiv.org/abs/2311.01599v1
- Date: Thu, 2 Nov 2023 21:10:16 GMT
- Title: Local Borsuk-Ulam, Stability, and Replicability
- Authors: Zachary Chase, Bogdan Chornomaz, Shay Moran, Amir Yehudayoff
- Abstract summary: We show that both list-replicable and globally stable learning are impossible in the PAC setting.
Specifically, we establish optimal bounds for list replicability and global stability numbers in finite classes.
- Score: 16.6959756920423
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We use and adapt the Borsuk-Ulam Theorem from topology to derive limitations
on list-replicable and globally stable learning algorithms. We further
demonstrate the applicability of our methods in combinatorics and topology.
We show that, besides trivial cases, both list-replicable and globally stable
learning are impossible in the agnostic PAC setting. This is in contrast with
the realizable case where it is known that any class with a finite Littlestone
dimension can be learned by such algorithms. In the realizable PAC setting, we
sharpen previous impossibility results and broaden their scope. Specifically,
we establish optimal bounds for list replicability and global stability numbers
in finite classes. This provides an exponential improvement over previous works
and implies an exponential separation from the Littlestone dimension. We
further introduce lower bounds for weak learners, i.e., learners that are only
marginally better than random guessing. Lower bounds from previous works apply
only to stronger learners.
To offer a broader and more comprehensive view of our topological approach,
we prove a local variant of the Borsuk-Ulam theorem in topology and a result in
combinatorics concerning Kneser colorings. In combinatorics, we prove that if
$c$ is a coloring of all non-empty subsets of $[n]$ such that disjoint sets
have different colors, then there is a chain of subsets that receives at least
$1+ \lfloor n/2\rfloor$ colors (this bound is sharp). In topology, we prove
e.g. that for any open antipodal-free cover of the $d$-dimensional sphere,
there is a point $x$ that belongs to at least $t=\lceil\frac{d+3}{2}\rceil$
sets.
Related papers
- Bridging the Gap Between Approximation and Learning via Optimal Approximation by ReLU MLPs of Maximal Regularity [8.28720658988688]
We identify a class of ReLU multilayer perceptions (MLPs) that are optimal function approximators and are statistically well-behaved.
We achieve this by avoiding the standard approach to constructing optimal ReLU approximators, which sacrifices by relying on small spikes.
arXiv Detail & Related papers (2024-09-18T22:05:07Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Prediction, Learning, Uniform Convergence, and Scale-sensitive
Dimensions [39.97534972432276]
We present a new general-purpose algorithm for learning classes of $[0,1]$-valued functions.
We prove a general upper bound on the expected absolute error of this algorithm.
We show how to apply both packing bounds to obtain improved general bounds on the sample complexity of learning.
arXiv Detail & Related papers (2023-04-21T15:51:35Z) - Pseudonorm Approachability and Applications to Regret Minimization [73.54127663296906]
We convert high-dimensional $ell_infty$-approachability problems to low-dimensional pseudonorm approachability problems.
We develop an algorithmic theory of pseudonorm approachability, analogous to previous work on approachability for $ell$ and other norms.
arXiv Detail & Related papers (2023-02-03T03:19:14Z) - Quantum and classical low-degree learning via a dimension-free Remez
inequality [52.12931955662553]
We show a new way to relate functions on the hypergrid to their harmonic extensions over the polytorus.
We show the supremum of a function $f$ over products of the cyclic group $exp(2pi i k/K)_k=1K$.
We extend to new spaces a recent line of work citeEI22, CHP, VZ22 that gave similarly efficient methods for learning low-degrees on hypercubes and observables on qubits.
arXiv Detail & Related papers (2023-01-04T04:15:40Z) - Agnostic Online Learning and Excellent Sets [21.87574015264402]
We prove existence of large indivisible'' sets in $k$-edge stable graphs.
A theme in these proofs is the interaction of two abstract notions of majority, arising from measure, and from rank or dimension.
arXiv Detail & Related papers (2021-08-12T07:33:33Z) - A simple geometric proof for the benefit of depth in ReLU networks [57.815699322370826]
We present a simple proof for the benefit of depth in multi-layer feedforward network with rectified activation ("depth separation")
We present a concrete neural network with linear depth (in $m$) and small constant width ($leq 4$) that classifies the problem with zero error.
arXiv Detail & Related papers (2021-01-18T15:40:27Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - Algorithms and SQ Lower Bounds for PAC Learning One-Hidden-Layer ReLU
Networks [48.32532049640782]
We study the problem of learning one-hidden-layer ReLU networks with $k$ hidden units on $mathbbRd$ under Gaussian marginals.
For the case of positive coefficients, we give the first-time algorithm for this learning problem for $k$ up to $tildeOOmega(sqrtlog d)$.
arXiv Detail & Related papers (2020-06-22T17:53:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.