Incremental Meta-Learning via Indirect Discriminant Alignment
- URL: http://arxiv.org/abs/2002.04162v2
- Date: Tue, 21 Apr 2020 18:19:18 GMT
- Title: Incremental Meta-Learning via Indirect Discriminant Alignment
- Authors: Qing Liu, Orchid Majumder, Alessandro Achille, Avinash Ravichandran,
Rahul Bhotika, Stefano Soatto
- Abstract summary: We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
- Score: 118.61152684795178
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Majority of the modern meta-learning methods for few-shot classification
tasks operate in two phases: a meta-training phase where the meta-learner
learns a generic representation by solving multiple few-shot tasks sampled from
a large dataset and a testing phase, where the meta-learner leverages its
learnt internal representation for a specific few-shot task involving classes
which were not seen during the meta-training phase. To the best of our
knowledge, all such meta-learning methods use a single base dataset for
meta-training to sample tasks from and do not adapt the algorithm after
meta-training. This strategy may not scale to real-world use-cases where the
meta-learner does not potentially have access to the full meta-training dataset
from the very beginning and we need to update the meta-learner in an
incremental fashion when additional training data becomes available. Through
our experimental setup, we develop a notion of incremental learning during the
meta-training phase of meta-learning and propose a method which can be used
with multiple existing metric-based meta-learning algorithms. Experimental
results on benchmark dataset show that our approach performs favorably at test
time as compared to training a model with the full meta-training set and incurs
negligible amount of catastrophic forgetting
Related papers
- Architecture, Dataset and Model-Scale Agnostic Data-free Meta-Learning [119.70303730341938]
We propose ePisode cUrriculum inveRsion (ECI) during data-free meta training and invErsion calibRation following inner loop (ICFIL) during meta testing.
ECI adaptively increases the difficulty level of pseudo episodes according to the real-time feedback of the meta model.
We formulate the optimization process of meta training with ECI as an adversarial form in an end-to-end manner.
arXiv Detail & Related papers (2023-03-20T15:10:41Z) - MetaICL: Learning to Learn In Context [87.23056864536613]
We introduce MetaICL, a new meta-training framework for few-shot learning where a pretrained language model is tuned to do in-context learn-ing on a large set of training tasks.
We show that MetaICL approaches (and sometimes beats) the performance of models fully finetuned on the target task training data, and outperforms much bigger models with nearly 8x parameters.
arXiv Detail & Related papers (2021-10-29T17:42:08Z) - Lessons from Chasing Few-Shot Learning Benchmarks: Rethinking the
Evaluation of Meta-Learning Methods [9.821362920940631]
We introduce a simple baseline for meta-learning, FIX-ML.
We explore two possible goals of meta-learning: to develop methods that generalize (i) to the same task distribution that generates the training set (in-distribution), or (ii) to new, unseen task distributions (out-of-distribution)
Our results highlight that in order to reason about progress in this space, it is necessary to provide a clearer description of the goals of meta-learning, and to develop more appropriate evaluation strategies.
arXiv Detail & Related papers (2021-02-23T05:34:30Z) - Variable-Shot Adaptation for Online Meta-Learning [123.47725004094472]
We study the problem of learning new tasks from a small, fixed number of examples, by meta-learning across static data from a set of previous tasks.
We find that meta-learning solves the full task set with fewer overall labels and greater cumulative performance, compared to standard supervised methods.
These results suggest that meta-learning is an important ingredient for building learning systems that continuously learn and improve over a sequence of problems.
arXiv Detail & Related papers (2020-12-14T18:05:24Z) - A Nested Bi-level Optimization Framework for Robust Few Shot Learning [10.147225934340877]
NestedMAML learns to assign weights to training tasks or instances.
Experiments on synthetic and real-world datasets demonstrate that NestedMAML efficiently mitigates the effects of "unwanted" tasks or instances.
arXiv Detail & Related papers (2020-11-13T06:41:22Z) - Improving Generalization in Meta-learning via Task Augmentation [69.83677015207527]
We propose two task augmentation methods, including MetaMix and Channel Shuffle.
Both MetaMix and Channel Shuffle outperform state-of-the-art results by a large margin across many datasets.
arXiv Detail & Related papers (2020-07-26T01:50:42Z) - Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning [79.25478727351604]
We explore a simple process: meta-learning over a whole-classification pre-trained model on its evaluation metric.
We observe this simple method achieves competitive performance to state-of-the-art methods on standard benchmarks.
arXiv Detail & Related papers (2020-03-09T20:06:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.