Does MAML Only Work via Feature Re-use? A Data Centric Perspective
- URL: http://arxiv.org/abs/2112.13137v1
- Date: Fri, 24 Dec 2021 20:18:38 GMT
- Title: Does MAML Only Work via Feature Re-use? A Data Centric Perspective
- Authors: Brando Miranda, Yu-Xiong Wang and Sanmi Koyejo
- Abstract summary: We provide empirical results that shed some light on how meta-learned MAML representations function.
We show that it is possible to define a family of synthetic benchmarks that result in a low degree of feature re-use.
We conjecture the core challenge of re-thinking meta-learning is in the design of few-shot learning data sets and benchmarks.
- Score: 19.556093984142418
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work has suggested that a good embedding is all we need to solve many
few-shot learning benchmarks. Furthermore, other work has strongly suggested
that Model Agnostic Meta-Learning (MAML) also works via this same method - by
learning a good embedding. These observations highlight our lack of
understanding of what meta-learning algorithms are doing and when they work. In
this work, we provide empirical results that shed some light on how
meta-learned MAML representations function. In particular, we identify three
interesting properties: 1) In contrast to previous work, we show that it is
possible to define a family of synthetic benchmarks that result in a low degree
of feature re-use - suggesting that current few-shot learning benchmarks might
not have the properties needed for the success of meta-learning algorithms; 2)
meta-overfitting occurs when the number of classes (or concepts) are finite,
and this issue disappears once the task has an unbounded number of concepts
(e.g., online learning); 3) more adaptation at meta-test time with MAML does
not necessarily result in a significant representation change or even an
improvement in meta-test performance - even when training on our proposed
synthetic benchmarks. Finally, we suggest that to understand meta-learning
algorithms better, we must go beyond tracking only absolute performance and, in
addition, formally quantify the degree of meta-learning and track both metrics
together. Reporting results in future work this way will help us identify the
sources of meta-overfitting more accurately and help us design more flexible
meta-learning algorithms that learn beyond fixed feature re-use. Finally, we
conjecture the core challenge of re-thinking meta-learning is in the design of
few-shot learning data sets and benchmarks - rather than in the algorithms, as
suggested by previous work.
Related papers
- Fast Adaptation with Kernel and Gradient based Meta Leaning [4.763682200721131]
We propose two algorithms to improve both the inner and outer loops of Model A Meta Learning (MAML)
Our first algorithm redefines the optimization problem in the function space to update the model using closed-form solutions.
In the outer loop, the second algorithm adjusts the learning of the meta-learner by assigning weights to the losses from each task of the inner loop.
arXiv Detail & Related papers (2024-11-01T07:05:03Z) - Context-Aware Meta-Learning [52.09326317432577]
We propose a meta-learning algorithm that emulates Large Language Models by learning new visual concepts during inference without fine-tuning.
Our approach exceeds or matches the state-of-the-art algorithm, P>M>F, on 8 out of 11 meta-learning benchmarks.
arXiv Detail & Related papers (2023-10-17T03:35:27Z) - Learning to Learn with Indispensable Connections [6.040904021861969]
We propose a novel meta-learning method called Meta-LTH that includes indispensible (necessary) connections.
Our method improves the classification accuracy by approximately 2% (20-way 1-shot task setting) for omniglot dataset.
arXiv Detail & Related papers (2023-04-06T04:53:13Z) - MAC: A Meta-Learning Approach for Feature Learning and Recombination [0.0]
In this paper, we invoke the width-depth duality of neural networks, wherein, we increase the width of the network by adding extra computational units (ACU)
Experimental results show that our proposed MAC method outperformed existing ANIL algorithm for non-similar task distribution by approximately 13%.
arXiv Detail & Related papers (2022-09-20T11:09:54Z) - What Matters For Meta-Learning Vision Regression Tasks? [19.373532562905208]
This paper makes two main contributions that help understand this barely explored area.
First, we design two new types of cross-category level vision regression tasks, namely object discovery and pose estimation.
Second, we propose the addition of functional contrastive learning (FCL) over the task representations in Conditional Neural Processes (CNPs) and train in an end-to-end fashion.
arXiv Detail & Related papers (2022-03-09T17:28:16Z) - Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and
Personalized Federated Learning [56.17603785248675]
Model-agnostic meta-learning (MAML) has become a popular research area.
Existing MAML algorithms rely on the episode' idea by sampling a few tasks and data points to update the meta-model at each iteration.
This paper proposes memory-based algorithms for MAML that converge with vanishing error.
arXiv Detail & Related papers (2021-06-09T08:47:58Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Rethinking Few-Shot Image Classification: a Good Embedding Is All You
Need? [72.00712736992618]
We show that a simple baseline: learning a supervised or self-supervised representation on the meta-training set, outperforms state-of-the-art few-shot learning methods.
An additional boost can be achieved through the use of self-distillation.
We believe that our findings motivate a rethinking of few-shot image classification benchmarks and the associated role of meta-learning algorithms.
arXiv Detail & Related papers (2020-03-25T17:58:42Z) - Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning [79.25478727351604]
We explore a simple process: meta-learning over a whole-classification pre-trained model on its evaluation metric.
We observe this simple method achieves competitive performance to state-of-the-art methods on standard benchmarks.
arXiv Detail & Related papers (2020-03-09T20:06:36Z) - Incremental Meta-Learning via Indirect Discriminant Alignment [118.61152684795178]
We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
arXiv Detail & Related papers (2020-02-11T01:39:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.