Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and
Unseen Tasks
- URL: http://arxiv.org/abs/2102.03832v1
- Date: Sun, 7 Feb 2021 16:16:23 GMT
- Title: Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and
Unseen Tasks
- Authors: Alireza Fallah, Aryan Mokhtari, Asuman Ozdaglar
- Abstract summary: We study the generalization properties of Model-Agnostic Meta-Learning (MAML) algorithms for supervised learning problems.
Our proof techniques rely on the connections between algorithmic stability and generalization bounds of algorithms.
- Score: 33.055672018805645
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we study the generalization properties of Model-Agnostic
Meta-Learning (MAML) algorithms for supervised learning problems. We focus on
the setting in which we train the MAML model over $m$ tasks, each with $n$ data
points, and characterize its generalization error from two points of view:
First, we assume the new task at test time is one of the training tasks, and we
show that, for strongly convex objective functions, the expected excess
population loss is bounded by $\mathcal{O}(1/mn)$. Second, we consider the MAML
algorithm's generalization to an unseen task and show that the resulting
generalization error depends on the total variation distance between the
underlying distributions of the new task and the tasks observed during the
training process. Our proof techniques rely on the connections between
algorithmic stability and generalization bounds of algorithms. In particular,
we propose a new definition of stability for meta-learning algorithms, which
allows us to capture the role of both the number of tasks $m$ and number of
samples per task $n$ on the generalization error of MAML.
Related papers
- On Discriminative Probabilistic Modeling for Self-Supervised Representation Learning [85.75164588939185]
We study the discriminative probabilistic modeling problem on a continuous domain for (multimodal) self-supervised representation learning.
We conduct generalization error analysis to reveal the limitation of current InfoNCE-based contrastive loss for self-supervised representation learning.
arXiv Detail & Related papers (2024-10-11T18:02:46Z) - Interpetable Target-Feature Aggregation for Multi-Task Learning based on Bias-Variance Analysis [53.38518232934096]
Multi-task learning (MTL) is a powerful machine learning paradigm designed to leverage shared knowledge across tasks to improve generalization and performance.
We propose an MTL approach at the intersection between task clustering and feature transformation based on a two-phase iterative aggregation of targets and features.
In both phases, a key aspect is to preserve the interpretability of the reduced targets and features through the aggregation with the mean, which is motivated by applications to Earth science.
arXiv Detail & Related papers (2024-06-12T08:30:16Z) - ST-MAML: A Stochastic-Task based Method for Task-Heterogeneous
Meta-Learning [12.215288736524268]
This paper proposes a novel method, ST-MAML, that empowers model-agnostic meta-learning (MAML) to learn from multiple task distributions.
We demonstrate that ST-MAML matches or outperforms the state-of-the-art on two few-shot image classification tasks, one curve regression benchmark, one image completion problem, and a real-world temperature prediction application.
arXiv Detail & Related papers (2021-09-27T18:54:50Z) - Learning-to-learn non-convex piecewise-Lipschitz functions [44.6133187924678]
We analyze the meta-learning of the algorithms for piecewise-Lipschitz functions, a non-task with applications to both machine learning algorithms.
We propose a practical meta-learning procedure that learns both the step-size of the algorithm from multiple online learning tasks.
arXiv Detail & Related papers (2021-08-19T16:22:48Z) - MAML is a Noisy Contrastive Learner [72.04430033118426]
Model-agnostic meta-learning (MAML) is one of the most popular and widely-adopted meta-learning algorithms nowadays.
We provide a new perspective to the working mechanism of MAML and discover that: MAML is analogous to a meta-learner using a supervised contrastive objective function.
We propose a simple but effective technique, zeroing trick, to alleviate such interference.
arXiv Detail & Related papers (2021-06-29T12:52:26Z) - Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and
Personalized Federated Learning [56.17603785248675]
Model-agnostic meta-learning (MAML) has become a popular research area.
Existing MAML algorithms rely on the episode' idea by sampling a few tasks and data points to update the meta-model at each iteration.
This paper proposes memory-based algorithms for MAML that converge with vanishing error.
arXiv Detail & Related papers (2021-06-09T08:47:58Z) - Meta-Learning with Fewer Tasks through Task Interpolation [67.03769747726666]
Current meta-learning algorithms require a large number of meta-training tasks, which may not be accessible in real-world scenarios.
By meta-learning with task gradient (MLTI), our approach effectively generates additional tasks by randomly sampling a pair of tasks and interpolating the corresponding features and labels.
Empirically, in our experiments on eight datasets from diverse domains, we find that the proposed general MLTI framework is compatible with representative meta-learning algorithms and consistently outperforms other state-of-the-art strategies.
arXiv Detail & Related papers (2021-06-04T20:15:34Z) - Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning [63.64636047748605]
We develop a new theoretical framework to provide convergence guarantee for the general multi-step MAML algorithm.
In particular, our results suggest that an inner-stage step needs to be chosen inversely proportional to $N$ of inner-stage steps in order for $N$ MAML to have guaranteed convergence.
arXiv Detail & Related papers (2020-02-18T19:17:54Z) - Task-Robust Model-Agnostic Meta-Learning [42.27488241647739]
We introduce the notion of "task-robustness" by reformulating the popular ModelAgnostic Meta-Learning (AML) objective.
The solution to this novel formulation is taskrobust in the sense that it places equal importance on even the most difficult/or rare tasks.
arXiv Detail & Related papers (2020-02-12T02:20:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.