Meta-Learning with Self-Improving Momentum Target
- URL: http://arxiv.org/abs/2210.05185v1
- Date: Tue, 11 Oct 2022 06:45:15 GMT
- Title: Meta-Learning with Self-Improving Momentum Target
- Authors: Jihoon Tack and Jongjin Park and Hankook Lee and Jaeho Lee and Jinwoo
Shin
- Abstract summary: We propose Self-improving Momentum Target (SiMT) to improve the performance of a meta-learner.
SiMT generates the target model by adapting from the temporal ensemble of the meta-learner.
We show that SiMT brings a significant performance gain when combined with a wide range of meta-learning methods.
- Score: 72.98879709228981
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The idea of using a separately trained target model (or teacher) to improve
the performance of the student model has been increasingly popular in various
machine learning domains, and meta-learning is no exception; a recent discovery
shows that utilizing task-wise target models can significantly boost the
generalization performance. However, obtaining a target model for each task can
be highly expensive, especially when the number of tasks for meta-learning is
large. To tackle this issue, we propose a simple yet effective method, coined
Self-improving Momentum Target (SiMT). SiMT generates the target model by
adapting from the temporal ensemble of the meta-learner, i.e., the momentum
network. This momentum network and its task-specific adaptations enjoy a
favorable generalization performance, enabling self-improving of the
meta-learner through knowledge distillation. Moreover, we found that perturbing
parameters of the meta-learner, e.g., dropout, further stabilize this
self-improving process by preventing fast convergence of the distillation loss
during meta-training. Our experimental results demonstrate that SiMT brings a
significant performance gain when combined with a wide range of meta-learning
methods under various applications, including few-shot regression, few-shot
classification, and meta-reinforcement learning. Code is available at
https://github.com/jihoontack/SiMT.
Related papers
- Unsupervised Representation Learning to Aid Semi-Supervised Meta
Learning [16.534014215010757]
We propose a one-shot unsupervised meta-learning to learn latent representation of training samples.
A temperature-scaled cross-entropy loss is used in the inner loop of meta-learning to prevent overfitting.
The proposed method is model agnostic and can aid any meta-learning model to improve accuracy.
arXiv Detail & Related papers (2023-10-19T18:25:22Z) - Improving Meta-Learning Generalization with Activation-Based
Early-Stopping [12.299371455015239]
Meta-Learning algorithms for few-shot learning aim to train neural networks capable of generalizing to novel tasks using only a few examples.
Early-stopping is critical for performance, halting model training when it reaches optimal generalization to the new task distribution.
This is problematic in few-shot transfer learning settings, where the meta-test set comes from a different target dataset.
arXiv Detail & Related papers (2022-08-03T22:55:45Z) - Energy-Efficient and Federated Meta-Learning via Projected Stochastic
Gradient Ascent [79.58680275615752]
We propose an energy-efficient federated meta-learning framework.
We assume each task is owned by a separate agent, so a limited number of tasks is used to train a meta-model.
arXiv Detail & Related papers (2021-05-31T08:15:44Z) - On Fast Adversarial Robustness Adaptation in Model-Agnostic
Meta-Learning [100.14809391594109]
Model-agnostic meta-learning (MAML) has emerged as one of the most successful meta-learning techniques in few-shot learning.
Despite the generalization power of the meta-model, it remains elusive that how adversarial robustness can be maintained by MAML in few-shot learning.
We propose a general but easily-optimized robustness-regularized meta-learning framework, which allows the use of unlabeled data augmentation, fast adversarial attack generation, and computationally-light fine-tuning.
arXiv Detail & Related papers (2021-02-20T22:03:04Z) - Generalising via Meta-Examples for Continual Learning in the Wild [24.09600678738403]
We develop a novel strategy to deal with neural networks that "learn in the wild"
We equip it with MEML - Meta-Example Meta-Learning - a new module that simultaneously alleviates catastrophic forgetting.
We extend it by adopting a technique that creates various augmented tasks and optimises over the hardest.
arXiv Detail & Related papers (2021-01-28T15:51:54Z) - Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning [79.25478727351604]
We explore a simple process: meta-learning over a whole-classification pre-trained model on its evaluation metric.
We observe this simple method achieves competitive performance to state-of-the-art methods on standard benchmarks.
arXiv Detail & Related papers (2020-03-09T20:06:36Z) - Curriculum in Gradient-Based Meta-Reinforcement Learning [10.447238563837173]
We show that gradient-based meta-learners are sensitive to task distributions.
With the wrong curriculum, agents suffer the effects of meta-overfitting, shallow adaptation, and adaptation instability.
arXiv Detail & Related papers (2020-02-19T01:40:45Z) - Unraveling Meta-Learning: Understanding Feature Representations for
Few-Shot Tasks [55.66438591090072]
We develop a better understanding of the underlying mechanics of meta-learning and the difference between models trained using meta-learning and models trained classically.
We develop a regularizer which boosts the performance of standard training routines for few-shot classification.
arXiv Detail & Related papers (2020-02-17T03:18:45Z) - Incremental Meta-Learning via Indirect Discriminant Alignment [118.61152684795178]
We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
arXiv Detail & Related papers (2020-02-11T01:39:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.