Distributional PAC-Learning from Nisan's Natural Proofs
- URL: http://arxiv.org/abs/2310.03641v2
- Date: Thu, 30 Nov 2023 21:01:06 GMT
- Title: Distributional PAC-Learning from Nisan's Natural Proofs
- Authors: Ari Karchmer
- Abstract summary: Natural proofs of circuit lower bounds for $Lambda imply efficient algorithms for learning $Lambda-circuits.
We consider whether this implication can be generalized to $Lambda notsupseteq AC0[p]$, and to learning algorithms which use only random examples and learn over arbitrary example distributions.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Carmosino et al. (2016) demonstrated that natural proofs of circuit lower
bounds for $\Lambda$ imply efficient algorithms for learning
$\Lambda$-circuits, but only over \textit{the uniform distribution}, with
\textit{membership queries}, and provided $\AC^0[p] \subseteq \Lambda$. We
consider whether this implication can be generalized to $\Lambda \not\supseteq
\AC^0[p]$, and to learning algorithms which use only random examples and learn
over arbitrary example distributions (Valiant's PAC-learning model).
We first observe that, if, for any circuit class $\Lambda$, there is an
implication from natural proofs for $\Lambda$ to PAC-learning for $\Lambda$,
then standard assumptions from lattice-based cryptography do not hold. In
particular, we observe that depth-2 majority circuits are a (conditional)
counter example to the implication, since Nisan (1993) gave a natural proof,
but Klivans and Sherstov (2009) showed hardness of PAC-learning under
lattice-based assumptions. We thus ask: what learning algorithms can we
reasonably expect to follow from Nisan's natural proofs?
Our main result is that all natural proofs arising from a type of
communication complexity argument, including Nisan's, imply PAC-learning
algorithms in a new \textit{distributional} variant (i.e., an ``average-case''
relaxation) of Valiant's PAC model. Our distributional PAC model is stronger
than the average-case prediction model of Blum et al. (1993) and the heuristic
PAC model of Nanashima (2021), and has several important properties which make
it of independent interest, such as being \textit{boosting-friendly}. The main
applications of our result are new distributional PAC-learning algorithms for
depth-2 majority circuits, polytopes and DNFs over natural target
distributions, as well as the nonexistence of encoded-input weak PRFs that can
be evaluated by depth-2 majority circuits.
Related papers
- The Sample Complexity of Smooth Boosting and the Tightness of the Hardcore Theorem [53.446980306786095]
Smooth boosters generate distributions that do not place too much weight on any given example.
Originally introduced for their noise-tolerant properties, such boosters have also found applications in differential privacy, mildly, and quantum learning theory.
arXiv Detail & Related papers (2024-09-17T23:09:25Z) - Fast Rates for Bandit PAC Multiclass Classification [73.17969992976501]
We study multiclass PAC learning with bandit feedback, where inputs are classified into one of $K$ possible labels and feedback is limited to whether or not the predicted labels are correct.
Our main contribution is in designing a novel learning algorithm for the agnostic $(varepsilon,delta)$PAC version of the problem.
arXiv Detail & Related papers (2024-06-18T08:54:04Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Attribute-Efficient PAC Learning of Low-Degree Polynomial Threshold Functions with Nasty Noise [10.550885570889527]
We study PAC learning of $K$sparse degree-$d$ PTFs on $mathbbRn$.
Our main contribution is a new algorithm that runs in time $(nd/epsilon)O(d)$.
PAC learns the class up to error rate $epsilon$ with $O(fracK4depsilon2d cdot log5d n)$ even when an $eta leq O(epsilond)
arXiv Detail & Related papers (2023-06-01T13:49:22Z) - Agnostic PAC Learning of k-juntas Using L2-Polynomial Regression [9.732863739456036]
We present a new PAC learning algorithm based on the Fourier expansion with lower computational complexity.
We prove our results by connecting the PAC learning with 0-1 loss to the minimum mean square estimation problem.
We derive an elegant upper bound on the 0-1 loss in terms of the MMSE error and show that the sign of the MMSE is a PAC learner for any concept class containing it.
arXiv Detail & Related papers (2023-03-08T19:54:07Z) - Differentially-Private Bayes Consistency [70.92545332158217]
We construct a Bayes consistent learning rule that satisfies differential privacy (DP)
We prove that any VC class can be privately learned in a semi-supervised setting with a near-optimal sample complexity.
arXiv Detail & Related papers (2022-12-08T11:57:30Z) - When are Local Queries Useful for Robust Learning? [25.832511407411637]
We study learning models where the learner is given more power through the use of local queries.
We give the first distribution-free algorithms that perform robust empirical risk minimization.
We finish by giving robust learning algorithms for halfspaces on $0,1n$ and then obtaining robustness guarantees for halfspaces in $mathbbRn$ against precision-bounded adversaries.
arXiv Detail & Related papers (2022-10-12T11:04:22Z) - Algorithms and SQ Lower Bounds for PAC Learning One-Hidden-Layer ReLU
Networks [48.32532049640782]
We study the problem of learning one-hidden-layer ReLU networks with $k$ hidden units on $mathbbRd$ under Gaussian marginals.
For the case of positive coefficients, we give the first-time algorithm for this learning problem for $k$ up to $tildeOOmega(sqrtlog d)$.
arXiv Detail & Related papers (2020-06-22T17:53:54Z) - Learning Halfspaces with Tsybakov Noise [50.659479930171585]
We study the learnability of halfspaces in the presence of Tsybakov noise.
We give an algorithm that achieves misclassification error $epsilon$ with respect to the true halfspace.
arXiv Detail & Related papers (2020-06-11T14:25:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.