Learning to Continually Learn
- URL: http://arxiv.org/abs/2002.09571v2
- Date: Wed, 4 Mar 2020 03:22:48 GMT
- Title: Learning to Continually Learn
- Authors: Shawn Beaulieu, Lapo Frati, Thomas Miconi, Joel Lehman, Kenneth O.
Stanley, Jeff Clune, Nick Cheney
- Abstract summary: We propose A Neuromodulated Meta-Learning Algorithm (ANML)
Inspired by neuromodulatory processes in the brain, we propose A Neuromodulated Meta-Learning Algorithm (ANML)
ANML produces state-of-the-art continual learning performance, sequentially learning as many as 600 classes (over 9,000 SGD updates)
- Score: 14.988129334830003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual lifelong learning requires an agent or model to learn many
sequentially ordered tasks, building on previous knowledge without
catastrophically forgetting it. Much work has gone towards preventing the
default tendency of machine learning models to catastrophically forget, yet
virtually all such work involves manually-designed solutions to the problem. We
instead advocate meta-learning a solution to catastrophic forgetting, allowing
AI to learn to continually learn. Inspired by neuromodulatory processes in the
brain, we propose A Neuromodulated Meta-Learning Algorithm (ANML). It
differentiates through a sequential learning process to meta-learn an
activation-gating function that enables context-dependent selective activation
within a deep neural network. Specifically, a neuromodulatory (NM) neural
network gates the forward pass of another (otherwise normal) neural network
called the prediction learning network (PLN). The NM network also thus
indirectly controls selective plasticity (i.e. the backward pass of) the PLN.
ANML enables continual learning without catastrophic forgetting at scale: it
produces state-of-the-art continual learning performance, sequentially learning
as many as 600 classes (over 9,000 SGD updates).
Related papers
- Neuro-mimetic Task-free Unsupervised Online Learning with Continual
Self-Organizing Maps [56.827895559823126]
Self-organizing map (SOM) is a neural model often used in clustering and dimensionality reduction.
We propose a generalization of the SOM, the continual SOM, which is capable of online unsupervised learning under a low memory budget.
Our results, on benchmarks including MNIST, Kuzushiji-MNIST, and Fashion-MNIST, show almost a two times increase in accuracy.
arXiv Detail & Related papers (2024-02-19T19:11:22Z) - Hebbian Learning based Orthogonal Projection for Continual Learning of
Spiking Neural Networks [74.3099028063756]
We develop a new method with neuronal operations based on lateral connections and Hebbian learning.
We show that Hebbian and anti-Hebbian learning on recurrent lateral connections can effectively extract the principal subspace of neural activities.
Our method consistently solves for spiking neural networks with nearly zero forgetting.
arXiv Detail & Related papers (2024-02-19T09:29:37Z) - Meta-Learning in Spiking Neural Networks with Reward-Modulated STDP [2.179313476241343]
We propose a bio-plausible meta-learning model inspired by the hippocampus and the prefrontal cortex.
Our new model can easily be applied to spike-based neuromorphic devices and enables fast learning in neuromorphic hardware.
arXiv Detail & Related papers (2023-06-07T13:08:46Z) - Neural Routing in Meta Learning [9.070747377130472]
We aim to improve the model performance of the current meta learning algorithms by selectively using only parts of the model conditioned on the input tasks.
In this work, we describe an approach that investigates task-dependent dynamic neuron selection in deep convolutional neural networks (CNNs) by leveraging the scaling factor in the batch normalization layer.
We find that the proposed approach, neural routing in meta learning (NRML), outperforms one of the well-known existing meta learning baselines on few-shot classification tasks.
arXiv Detail & Related papers (2022-10-14T16:31:24Z) - Learning to learn online with neuromodulated synaptic plasticity in
spiking neural networks [0.0]
We show that models of neuromodulated synaptic plasticity from neuroscience can be trained to learn through gradient descent.
This framework opens a new path toward developing neuroscience inspired online learning algorithms.
arXiv Detail & Related papers (2022-06-25T00:28:40Z) - aSTDP: A More Biologically Plausible Learning [0.0]
We introduce approximate STDP, a new neural networks learning framework.
It uses only STDP rules for supervised and unsupervised learning.
It can make predictions or generate patterns in one model without additional configuration.
arXiv Detail & Related papers (2022-05-22T08:12:50Z) - Learning Bayesian Sparse Networks with Full Experience Replay for
Continual Learning [54.7584721943286]
Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered.
Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal.
We propose to only activate and select sparse neurons for learning current and past tasks at any stage.
arXiv Detail & Related papers (2022-02-21T13:25:03Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Artificial Neural Variability for Deep Learning: On Overfitting, Noise
Memorization, and Catastrophic Forgetting [135.0863818867184]
artificial neural variability (ANV) helps artificial neural networks learn some advantages from natural'' neural networks.
ANV plays as an implicit regularizer of the mutual information between the training data and the learned model.
It can effectively relieve overfitting, label noise memorization, and catastrophic forgetting at negligible costs.
arXiv Detail & Related papers (2020-11-12T06:06:33Z) - Online Few-shot Gesture Learning on a Neuromorphic Processor [9.084047904714629]
We present the Surrogate-gradient Online Error-triggered Learning (SOEL) system for online few-shot learning on neuromorphic processors.
SOEL updates trigger when an error occurs, enabling faster learning with fewer updates.
arXiv Detail & Related papers (2020-08-03T19:39:54Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.