User-friendly introduction to PAC-Bayes bounds
- URL: http://arxiv.org/abs/2110.11216v1
- Date: Thu, 21 Oct 2021 15:50:05 GMT
- Title: User-friendly introduction to PAC-Bayes bounds
- Authors: Pierre Alquier
- Abstract summary: In statistical learning theory there is a set of tools designed to understand the generalization ability of procedures: PAC-Bayesian or PAC-Bayes bounds.
Very recently, PAC-Bayes bounds received a considerable attention: for example there was workshop on PAC-Bayesian trends and insights", organized by B. Guedj, F. Bach and P. Germain.
- Score: 0.6599344783327052
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Aggregated predictors are obtained by making a set of basic predictors vote
according to some weights, that is, to some probability distribution.
Randomized predictors are obtained by sampling in a set of basic predictors,
according to some prescribed probability distribution.
Thus, aggregated and randomized predictors have in common that they are not
defined by a minimization problem, but by a probability distribution on the set
of predictors. In statistical learning theory, there is a set of tools designed
to understand the generalization ability of such procedures: PAC-Bayesian or
PAC-Bayes bounds.
Since the original PAC-Bayes bounds of McAllester, these tools have been
considerably improved in many directions (we will for example describe a
simplified version of the localization technique of Catoni that was missed by
the community, and later rediscovered as "mutual information bounds"). Very
recently, PAC-Bayes bounds received a considerable attention: for example there
was workshop on PAC-Bayes at NIPS 2017, "(Almost) 50 Shades of Bayesian
Learning: PAC-Bayesian trends and insights", organized by B. Guedj, F. Bach and
P. Germain. One of the reason of this recent success is the successful
application of these bounds to neural networks by Dziugaite and Roy.
An elementary introduction to PAC-Bayes theory is still missing. This is an
attempt to provide such an introduction.
Related papers
- PAC Prediction Sets Under Label Shift [52.30074177997787]
Prediction sets capture uncertainty by predicting sets of labels rather than individual labels.
We propose a novel algorithm for constructing prediction sets with PAC guarantees in the label shift setting.
We evaluate our approach on five datasets.
arXiv Detail & Related papers (2023-10-19T17:57:57Z) - Variational Prediction [95.00085314353436]
We present a technique for learning a variational approximation to the posterior predictive distribution using a variational bound.
This approach can provide good predictive distributions without test time marginalization costs.
arXiv Detail & Related papers (2023-07-14T18:19:31Z) - ZigZag: Universal Sampling-free Uncertainty Estimation Through Two-Step Inference [54.17205151960878]
We introduce a sampling-free approach that is generic and easy to deploy.
We produce reliable uncertainty estimates on par with state-of-the-art methods at a significantly lower computational cost.
arXiv Detail & Related papers (2022-11-21T13:23:09Z) - Analyzing Lottery Ticket Hypothesis from PAC-Bayesian Theory Perspective [25.157282476221482]
We show that the PAC-Bayesian theory can provide an explicit understanding of the relationship between LTH and generalization behavior.
We offer the PAC-Bayes bound using a spike-and-slab distribution to analyze winning tickets.
arXiv Detail & Related papers (2022-05-15T15:58:27Z) - Robust PAC$^m$: Training Ensemble Models Under Misspecification and
Outliers [46.38465729190199]
PAC-Bayes theory demonstrates that the free energy criterion minimized by Bayesian learning is a bound on the generalization error for Gibbs predictors.
This work presents a novel robust free energy criterion that combines the generalized score function with PAC$m$ ensemble bounds.
arXiv Detail & Related papers (2022-03-03T17:11:07Z) - A General Framework for the Practical Disintegration of PAC-Bayesian
Bounds [2.516393111664279]
We introduce new PAC-Bayesian generalization bounds that have the originality to provide disintegrated bounds.
Our bounds are easily optimizable and can be used to design learning algorithms.
arXiv Detail & Related papers (2021-02-17T09:36:46Z) - A PAC-Bayesian Perspective on Structured Prediction with Implicit Loss
Embeddings [11.711761707845865]
PAC-Bayes has gained interest recently for its capacity of producing tight risk bounds for predictor distributions.
We present two generalization bounds, on the risk and excess risk, which yield insights into the behavior of ILE predictors.
Two learning algorithms are derived from these bounds.
arXiv Detail & Related papers (2020-12-07T15:19:43Z) - PAC$^m$-Bayes: Narrowing the Empirical Risk Gap in the Misspecified
Bayesian Regime [75.19403612525811]
This work develops a multi-sample loss which can close the gap by spanning a trade-off between the two risks.
Empirical study demonstrates improvement to the predictive distribution.
arXiv Detail & Related papers (2020-10-19T16:08:34Z) - PAC-Bayes Analysis Beyond the Usual Bounds [16.76187007910588]
We focus on a learning model where the learner observes a finite set of training examples.
The learned data-dependent distribution is then used to make randomized predictions.
arXiv Detail & Related papers (2020-06-23T14:30:24Z) - Distributionally Robust Bayesian Quadrature Optimization [60.383252534861136]
We study BQO under distributional uncertainty in which the underlying probability distribution is unknown except for a limited set of its i.i.d. samples.
A standard BQO approach maximizes the Monte Carlo estimate of the true expected objective given the fixed sample set.
We propose a novel posterior sampling based algorithm, namely distributionally robust BQO (DRBQO) for this purpose.
arXiv Detail & Related papers (2020-01-19T12:00:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.