PAC-Bayes Bounds for Meta-learning with Data-Dependent Prior
- URL: http://arxiv.org/abs/2102.03748v1
- Date: Sun, 7 Feb 2021 09:03:43 GMT
- Title: PAC-Bayes Bounds for Meta-learning with Data-Dependent Prior
- Authors: Tianyu Liu, Jie Lu, Zheng Yan, Guangquan Zhang
- Abstract summary: We derive three novel generalisation error bounds for meta-learning based on PAC-Bayes relative entropy bound.
Experiments illustrate that the proposed three PAC-Bayes bounds for meta-learning guarantee a competitive generalization performance guarantee.
- Score: 36.38937352131301
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: By leveraging experience from previous tasks, meta-learning algorithms can
achieve effective fast adaptation ability when encountering new tasks. However
it is unclear how the generalization property applies to new tasks. Probably
approximately correct (PAC) Bayes bound theory provides a theoretical framework
to analyze the generalization performance for meta-learning. We derive three
novel generalisation error bounds for meta-learning based on PAC-Bayes relative
entropy bound. Furthermore, using the empirical risk minimization (ERM) method,
a PAC-Bayes bound for meta-learning with data-dependent prior is developed.
Experiments illustrate that the proposed three PAC-Bayes bounds for
meta-learning guarantee a competitive generalization performance guarantee, and
the extended PAC-Bayes bound with data-dependent prior can achieve rapid
convergence ability.
Related papers
- Improving Generalization of Complex Models under Unbounded Loss Using PAC-Bayes Bounds [10.94126149188336]
PAC-Bayes learning theory has focused extensively on establishing tight upper bounds for test errors.
A recently proposed training procedure called PAC-Bayes training, updates the model toward minimizing these bounds.
This approach is theoretically sound, in practice, it has not achieved a test error as low as those obtained by empirical risk minimization (ERM)
We introduce a new PAC-Bayes training algorithm with improved performance and reduced reliance on prior tuning.
arXiv Detail & Related papers (2023-05-30T17:31:25Z) - Scalable PAC-Bayesian Meta-Learning via the PAC-Optimal Hyper-Posterior:
From Theory to Practice [54.03076395748459]
A central question in the meta-learning literature is how to regularize to ensure generalization to unseen tasks.
We present a generalization bound for meta-learning, which was first derived by Rothfuss et al.
We provide a theoretical analysis and empirical case study under which conditions and to what extent these guarantees for meta-learning improve upon PAC-Bayesian per-task learning bounds.
arXiv Detail & Related papers (2022-11-14T08:51:04Z) - A General framework for PAC-Bayes Bounds for Meta-Learning [0.0]
We study PAC-Bayes bounds on meta generalization gap.
In this paper, by upper bounding arbitrary convex functions, we obtain new PAC-Bayes bounds.
Using these bounds, we develop new PAC-Bayes meta-learning algorithms.
arXiv Detail & Related papers (2022-06-11T07:45:25Z) - Towards Scaling Difference Target Propagation by Learning Backprop
Targets [64.90165892557776]
Difference Target Propagation is a biologically-plausible learning algorithm with close relation with Gauss-Newton (GN) optimization.
We propose a novel feedback weight training scheme that ensures both that DTP approximates BP and that layer-wise feedback weight training can be restored.
We report the best performance ever achieved by DTP on CIFAR-10 and ImageNet.
arXiv Detail & Related papers (2022-01-31T18:20:43Z) - APS: Active Pretraining with Successor Features [96.24533716878055]
We show that by reinterpreting and combining successorcitepHansenFast with non entropy, the intractable mutual information can be efficiently optimized.
The proposed method Active Pretraining with Successor Feature (APS) explores the environment via non entropy, and the explored data can be efficiently leveraged to learn behavior.
arXiv Detail & Related papers (2021-08-31T16:30:35Z) - Meta-Learning with Fewer Tasks through Task Interpolation [67.03769747726666]
Current meta-learning algorithms require a large number of meta-training tasks, which may not be accessible in real-world scenarios.
By meta-learning with task gradient (MLTI), our approach effectively generates additional tasks by randomly sampling a pair of tasks and interpolating the corresponding features and labels.
Empirically, in our experiments on eight datasets from diverse domains, we find that the proposed general MLTI framework is compatible with representative meta-learning algorithms and consistently outperforms other state-of-the-art strategies.
arXiv Detail & Related papers (2021-06-04T20:15:34Z) - PAC-BUS: Meta-Learning Bounds via PAC-Bayes and Uniform Stability [3.42658286826597]
We derive a probably correct (PAC) bound for stable-based meta-learning using two different generalization levels.
We present a practical regularization scheme motivated by the bound in settings where the bound is at the baseline level.
arXiv Detail & Related papers (2021-02-12T15:57:45Z) - PACOH: Bayes-Optimal Meta-Learning with PAC-Guarantees [77.67258935234403]
We provide a theoretical analysis using the PAC-Bayesian framework and derive novel generalization bounds for meta-learning.
We develop a class of PAC-optimal meta-learning algorithms with performance guarantees and a principled meta-level regularization.
arXiv Detail & Related papers (2020-02-13T15:01:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.