A Characterization of List Learnability
- URL: http://arxiv.org/abs/2211.04956v2
- Date: Sun, 26 Mar 2023 08:17:43 GMT
- Title: A Characterization of List Learnability
- Authors: Moses Charikar, Chirag Pabbaraju
- Abstract summary: We consider list PAC learning where the goal is to output a list of $k$ predictions.
Generalizing the recent characterization of multiclass learnability, we show that a hypothesis class is $k$-list learnable if and only if the $k$-DS dimension is finite.
- Score: 15.368858716555888
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A classical result in learning theory shows the equivalence of PAC
learnability of binary hypothesis classes and the finiteness of VC dimension.
Extending this to the multiclass setting was an open problem, which was settled
in a recent breakthrough result characterizing multiclass PAC learnability via
the DS dimension introduced earlier by Daniely and Shalev-Shwartz. In this work
we consider list PAC learning where the goal is to output a list of $k$
predictions. List learning algorithms have been developed in several settings
before and indeed, list learning played an important role in the recent
characterization of multiclass learnability. In this work we ask: when is it
possible to $k$-list learn a hypothesis class? We completely characterize
$k$-list learnability in terms of a generalization of DS dimension that we call
the $k$-DS dimension. Generalizing the recent characterization of multiclass
learnability, we show that a hypothesis class is $k$-list learnable if and only
if the $k$-DS dimension is finite.
Related papers
- A Characterization of List Regression [5.371337604556312]
We provide a complete characterization of list regression.
We propose two dimensions, namely the $k$-OIG dimension and the $k$-fat-shattering dimension, and show that they optimally characterize realizable and $k$-list regression respectively.
arXiv Detail & Related papers (2024-09-28T03:19:37Z) - Beyond Prototypes: Semantic Anchor Regularization for Better
Representation Learning [82.29761875805369]
One of the ultimate goals of representation learning is to achieve compactness within a class and well-separability between classes.
We propose a novel perspective to use pre-defined class anchors serving as feature centroid to unidirectionally guide feature learning.
The proposed Semantic Anchor Regularization (SAR) can be used in a plug-and-play manner in the existing models.
arXiv Detail & Related papers (2023-12-19T05:52:38Z) - When is Agnostic Reinforcement Learning Statistically Tractable? [76.1408672715773]
A new complexity measure, called the emphspanning capacity, depends solely on the set $Pi$ and is independent of the MDP dynamics.
We show there exists a policy class $Pi$ with a bounded spanning capacity that requires a superpolynomial number of samples to learn.
This reveals a surprising separation for learnability between generative access and online access models.
arXiv Detail & Related papers (2023-10-09T19:40:54Z) - Multiclass Boosting: Simple and Intuitive Weak Learning Criteria [72.71096438538254]
We give a simple and efficient boosting algorithm, that does not require realizability assumptions.
We present a new result on boosting for list learners, as well as provide a novel proof for the characterization of multiclass PAC learning.
arXiv Detail & Related papers (2023-07-02T19:26:58Z) - Ticketed Learning-Unlearning Schemes [57.89421552780526]
We propose a new ticketed model for learning--unlearning.
We provide space-efficient ticketed learning--unlearning schemes for a broad family of concept classes.
arXiv Detail & Related papers (2023-06-27T18:54:40Z) - Comparative Learning: A Sample Complexity Theory for Two Hypothesis
Classes [5.194264506657145]
We introduce comparative learning as a combination of the realizable and agnostic settings in PAC learning.
Even when both $S$ and $B$ have infinite VC dimensions, comparative learning can still have a small sample complexity.
We show that the sample complexity of comparative learning is characterized by the mutual VC dimension $mathsfVC(S,B)$.
arXiv Detail & Related papers (2022-11-16T18:38:24Z) - Multiclass Learnability Beyond the PAC Framework: Universal Rates and
Partial Concept Classes [31.2676304636432]
We study the problem of multiclass classification with a bounded number of different labels $k$, in the realizable setting.
We extend the traditional PAC model to a) distribution-dependent learning rates, and b) learning rates under data-dependent assumptions.
arXiv Detail & Related papers (2022-10-05T14:36:27Z) - A Characterization of Multiclass Learnability [18.38631912121182]
We characterize multiclass PAC learnability through the DS dimension, a dimension defined by Daniely and Shalev-Shwartz 2014.
In the list learning setting, instead of predicting a single outcome for a given unseen input, the goal is to provide a short menu of predictions.
Our second main result concerns the Natarajan dimension, which has been a central candidate for characterizing multiclass learnability.
arXiv Detail & Related papers (2022-03-03T07:41:54Z) - Sample-efficient proper PAC learning with approximate differential
privacy [51.09425023771246]
We prove that the sample complexity of properly learning a class of Littlestone dimension $d$ with approximate differential privacy is $tilde O(d6)$, ignoring privacy and accuracy parameters.
This result answers a question of Bun et al. (FOCS 2020) by improving upon their upper bound of $2O(d)$ on the sample complexity.
arXiv Detail & Related papers (2020-12-07T18:17:46Z) - The Information Bottleneck Problem and Its Applications in Machine
Learning [53.57797720793437]
Inference capabilities of machine learning systems skyrocketed in recent years, now playing a pivotal role in various aspect of society.
The information bottleneck (IB) theory emerged as a bold information-theoretic paradigm for analyzing deep learning (DL) systems.
In this tutorial we survey the information-theoretic origins of this abstract principle, and its recent impact on DL.
arXiv Detail & Related papers (2020-04-30T16:48:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.