Generalising via Meta-Examples for Continual Learning in the Wild
- URL: http://arxiv.org/abs/2101.12081v1
- Date: Thu, 28 Jan 2021 15:51:54 GMT
- Title: Generalising via Meta-Examples for Continual Learning in the Wild
- Authors: Alessia Bertugli, Stefano Vincenzi, Simone Calderara, Andrea Passerini
- Abstract summary: We develop a novel strategy to deal with neural networks that "learn in the wild"
We equip it with MEML - Meta-Example Meta-Learning - a new module that simultaneously alleviates catastrophic forgetting.
We extend it by adopting a technique that creates various augmented tasks and optimises over the hardest.
- Score: 24.09600678738403
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning quickly and continually is still an ambitious task for neural
networks. Indeed, many real-world applications do not reflect the learning
setting where neural networks shine, as data are usually few, mostly unlabelled
and come as a stream. To narrow this gap, we introduce FUSION - Few-shot
UnSupervIsed cONtinual learning - a novel strategy which aims to deal with
neural networks that "learn in the wild", simulating a real distribution and
flow of unbalanced tasks. We equip FUSION with MEML - Meta-Example
Meta-Learning - a new module that simultaneously alleviates catastrophic
forgetting and favours the generalisation and future learning of new tasks. To
encourage features reuse during the meta-optimisation, our model exploits a
single inner loop per task, taking advantage of an aggregated representation
achieved through the use of a self-attention mechanism. To further enhance the
generalisation capability of MEML, we extend it by adopting a technique that
creates various augmented tasks and optimises over the hardest. Experimental
results on few-shot learning benchmarks show that our model exceeds the other
baselines in both FUSION and fully supervised case. We also explore how it
behaves in standard continual learning consistently outperforming
state-of-the-art approaches.
Related papers
- ConML: A Universal Meta-Learning Framework with Task-Level Contrastive Learning [49.447777286862994]
ConML is a universal meta-learning framework that can be applied to various meta-learning algorithms.
We demonstrate that ConML integrates seamlessly with optimization-based, metric-based, and amortization-based meta-learning algorithms.
arXiv Detail & Related papers (2024-10-08T12:22:10Z) - A Practitioner's Guide to Continual Multimodal Pretraining [83.63894495064855]
Multimodal foundation models serve numerous applications at the intersection of vision and language.
To keep models updated, research into continual pretraining mainly explores scenarios with either infrequent, indiscriminate updates on large-scale new data, or frequent, sample-level updates.
We introduce FoMo-in-Flux, a continual multimodal pretraining benchmark with realistic compute constraints and practical deployment requirements.
arXiv Detail & Related papers (2024-08-26T17:59:01Z) - Neuro-mimetic Task-free Unsupervised Online Learning with Continual
Self-Organizing Maps [56.827895559823126]
Self-organizing map (SOM) is a neural model often used in clustering and dimensionality reduction.
We propose a generalization of the SOM, the continual SOM, which is capable of online unsupervised learning under a low memory budget.
Our results, on benchmarks including MNIST, Kuzushiji-MNIST, and Fashion-MNIST, show almost a two times increase in accuracy.
arXiv Detail & Related papers (2024-02-19T19:11:22Z) - Learning Universal Predictors [23.18743879588599]
We explore the potential of amortizing the most powerful universal predictor, namely Solomonoff Induction (SI), into neural networks via leveraging meta-learning to its limits.
We use Universal Turing Machines (UTMs) to generate training data used to expose networks to a broad range of patterns.
Our results suggest that UTM data is a valuable resource for meta-learning, and that it can be used to train neural networks capable of learning universal prediction strategies.
arXiv Detail & Related papers (2024-01-26T15:37:16Z) - Predictive Experience Replay for Continual Visual Control and
Forecasting [62.06183102362871]
We present a new continual learning approach for visual dynamics modeling and explore its efficacy in visual control and forecasting.
We first propose the mixture world model that learns task-specific dynamics priors with a mixture of Gaussians, and then introduce a new training strategy to overcome catastrophic forgetting.
Our model remarkably outperforms the naive combinations of existing continual learning and visual RL algorithms on DeepMind Control and Meta-World benchmarks with continual visual control tasks.
arXiv Detail & Related papers (2023-03-12T05:08:03Z) - Meta-Learning via Classifier(-free) Guidance [5.812784742024491]
State-of-the-art meta-learning techniques do not optimize for zero-shot adaptation to unseen tasks.
We propose meta-learning techniques that use natural language guidance to achieve higher zero-shot performance.
arXiv Detail & Related papers (2022-10-17T11:09:35Z) - Meta-Learning with Self-Improving Momentum Target [72.98879709228981]
We propose Self-improving Momentum Target (SiMT) to improve the performance of a meta-learner.
SiMT generates the target model by adapting from the temporal ensemble of the meta-learner.
We show that SiMT brings a significant performance gain when combined with a wide range of meta-learning methods.
arXiv Detail & Related papers (2022-10-11T06:45:15Z) - Online Fast Adaptation and Knowledge Accumulation: a New Approach to
Continual Learning [74.07455280246212]
Continual learning studies agents that learn from streams of tasks without forgetting previous ones while adapting to new ones.
We show that current continual learning, meta-learning, meta-continual learning, and continual-meta learning techniques fail in this new scenario.
We propose Continual-MAML, an online extension of the popular MAML algorithm as a strong baseline for this scenario.
arXiv Detail & Related papers (2020-03-12T15:47:16Z) - Is the Meta-Learning Idea Able to Improve the Generalization of Deep
Neural Networks on the Standard Supervised Learning? [34.00378876525579]
We propose a novel metalearning based training procedure (M) for deep neural networks (DNNs)
M simulates the meta-training process by considering a batch of training samples as a task.
The experimental results show a consistently improved performance on all the generalizations with different sizes.
arXiv Detail & Related papers (2020-02-27T21:29:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.