PAC Verification of Statistical Algorithms
- URL: http://arxiv.org/abs/2211.17096v2
- Date: Sat, 2 Sep 2023 21:53:42 GMT
- Title: PAC Verification of Statistical Algorithms
- Authors: Saachi Mutreja, Jonathan Shafer
- Abstract summary: We develop the setting of PAC verification, where a hypothesis (machine learning model) that purportedly satisfies the agnostic PAC learning objective is verified using an interactive proof.
We present a protocol for PAC verification of unions of intervals over $mathbbR$ that improves upon their proposed protocol for that task, and matches our lower bound's dependence on $d$.
Our final result is a protocol for the verification of statistical algorithms that satisfy a constraint on their queries.
- Score: 0.10878040851637999
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Goldwasser et al. (2021) recently proposed the setting of PAC verification,
where a hypothesis (machine learning model) that purportedly satisfies the
agnostic PAC learning objective is verified using an interactive proof. In this
paper we develop this notion further in a number of ways. First, we prove a
lower bound of $\Omega\left(\sqrt{d}/\varepsilon^2\right)$ i.i.d.\ samples for
PAC verification of hypothesis classes of VC dimension $d$. Second, we present
a protocol for PAC verification of unions of intervals over $\mathbb{R}$ that
improves upon their proposed protocol for that task, and matches our lower
bound's dependence on $d$. Third, we introduce a natural generalization of
their definition to verification of general statistical algorithms, which is
applicable to a wider variety of settings beyond agnostic PAC learning.
Showcasing our proposed definition, our final result is a protocol for the
verification of statistical query algorithms that satisfy a combinatorial
constraint on their queries.
Related papers
- Efficient Learning of POMDPs with Known Observation Model in Average-Reward Setting [56.92178753201331]
We propose the Observation-Aware Spectral (OAS) estimation technique, which enables the POMDP parameters to be learned from samples collected using a belief-based policy.
We show the consistency of the OAS procedure, and we prove a regret guarantee of order $mathcalO(sqrtT log(T)$ for the proposed OAS-UCRL algorithm.
arXiv Detail & Related papers (2024-10-02T08:46:34Z) - Fast Rates for Bandit PAC Multiclass Classification [73.17969992976501]
We study multiclass PAC learning with bandit feedback, where inputs are classified into one of $K$ possible labels and feedback is limited to whether or not the predicted labels are correct.
Our main contribution is in designing a novel learning algorithm for the agnostic $(varepsilon,delta)$PAC version of the problem.
arXiv Detail & Related papers (2024-06-18T08:54:04Z) - Uniform Generalization Bounds on Data-Dependent Hypothesis Sets via PAC-Bayesian Theory on Random Sets [25.250314934981233]
We first apply the PAC-Bayesian framework on random sets' in a rigorous way, where the training algorithm is assumed to output a data-dependent hypothesis set.
This approach allows us to prove data-dependent bounds, which can be applicable in numerous contexts.
arXiv Detail & Related papers (2024-04-26T14:28:18Z) - Private Truly-Everlasting Robust-Prediction [22.662255469977598]
Private Everlasting Prediction (PEP) is a model for differentially private learning in which the learner never publicly releases a hypothesis.
PEP provides privacy both for the initial training set and for the endless stream of classification queries.
We present two conceptual modifications to the definition of PEP, as well as new constructions exhibiting significant improvements over prior work.
arXiv Detail & Related papers (2024-01-09T02:00:55Z) - Distributional PAC-Learning from Nisan's Natural Proofs [0.0]
Natural proofs of circuit lower bounds for $Lambda imply efficient algorithms for learning $Lambda-circuits.
We consider whether this implication can be generalized to $Lambda notsupseteq AC0[p]$, and to learning algorithms which use only random examples and learn over arbitrary example distributions.
arXiv Detail & Related papers (2023-10-05T16:13:29Z) - SHAP@k:Efficient and Probably Approximately Correct (PAC) Identification
of Top-k Features [16.99004256148679]
We introduce the Top-k Identification Problem (TkIP), where the objective is to identify the k features with the highest SHAP values.
The goal of our work is to improve the sample efficiency of existing methods in the context of solving TkIP.
arXiv Detail & Related papers (2023-07-10T18:42:45Z) - Learning versus Refutation in Noninteractive Local Differential Privacy [133.80204506727526]
We study two basic statistical tasks in non-interactive local differential privacy (LDP): learning and refutation.
Our main result is a complete characterization of the sample complexity of PAC learning for non-interactive LDP protocols.
arXiv Detail & Related papers (2022-10-26T03:19:24Z) - Nearly Optimal Latent State Decoding in Block MDPs [74.51224067640717]
In episodic Block MDPs, the decision maker has access to rich observations or contexts generated from a small number of latent states.
We are first interested in estimating the latent state decoding function based on data generated under a fixed behavior policy.
We then study the problem of learning near-optimal policies in the reward-free framework.
arXiv Detail & Related papers (2022-08-17T18:49:53Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - A Limitation of the PAC-Bayes Framework [32.24251308425503]
We present a limitation for the PAC-Bayes framework.
We demonstrate an easy learning task that is not amenable to a PAC-Bayes analysis.
arXiv Detail & Related papers (2020-06-24T06:36:00Z) - Breaking the Sample Size Barrier in Model-Based Reinforcement Learning
with a Generative Model [50.38446482252857]
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator)
We first consider $gamma$-discounted infinite-horizon Markov decision processes (MDPs) with state space $mathcalS$ and action space $mathcalA$.
We prove that a plain model-based planning algorithm suffices to achieve minimax-optimal sample complexity given any target accuracy level.
arXiv Detail & Related papers (2020-05-26T17:53:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.