Bio-Inspired, Task-Free Continual Learning through Activity
Regularization
- URL: http://arxiv.org/abs/2212.04316v1
- Date: Thu, 8 Dec 2022 15:14:20 GMT
- Title: Bio-Inspired, Task-Free Continual Learning through Activity
Regularization
- Authors: Francesco L\"assig, Pau Vilimelis Aceituno, Martino Sorbaro, Benjamin
F. Grewe
- Abstract summary: Continual learning approaches usually require discrete task boundaries.
We take inspiration from neuroscience, where sparse, non-overlapping neuronal representations have been suggested to prevent forgetting.
In addition to sparsity, we introduce lateral recurrent connections within each layer to further protect previously learned representations.
Our method achieves similar performance to well-known CL methods, such as Elastic Weight Consolidation and Synaptic Intelligence, without requiring information about task boundaries.
- Score: 3.5502600490147196
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The ability to sequentially learn multiple tasks without forgetting is a key
skill of biological brains, whereas it represents a major challenge to the
field of deep learning. To avoid catastrophic forgetting, various continual
learning (CL) approaches have been devised. However, these usually require
discrete task boundaries. This requirement seems biologically implausible and
often limits the application of CL methods in the real world where tasks are
not always well defined. Here, we take inspiration from neuroscience, where
sparse, non-overlapping neuronal representations have been suggested to prevent
catastrophic forgetting. As in the brain, we argue that these sparse
representations should be chosen on the basis of feed forward
(stimulus-specific) as well as top-down (context-specific) information. To
implement such selective sparsity, we use a bio-plausible form of hierarchical
credit assignment known as Deep Feedback Control (DFC) and combine it with a
winner-take-all sparsity mechanism. In addition to sparsity, we introduce
lateral recurrent connections within each layer to further protect previously
learned representations. We evaluate the new sparse-recurrent version of DFC on
the split-MNIST computer vision benchmark and show that only the combination of
sparsity and intra-layer recurrent connections improves CL performance with
respect to standard backpropagation. Our method achieves similar performance to
well-known CL methods, such as Elastic Weight Consolidation and Synaptic
Intelligence, without requiring information about task boundaries. Overall, we
showcase the idea of adopting computational principles from the brain to derive
new, task-free learning algorithms for CL.
Related papers
- Neuro-mimetic Task-free Unsupervised Online Learning with Continual
Self-Organizing Maps [56.827895559823126]
Self-organizing map (SOM) is a neural model often used in clustering and dimensionality reduction.
We propose a generalization of the SOM, the continual SOM, which is capable of online unsupervised learning under a low memory budget.
Our results, on benchmarks including MNIST, Kuzushiji-MNIST, and Fashion-MNIST, show almost a two times increase in accuracy.
arXiv Detail & Related papers (2024-02-19T19:11:22Z) - Fully Spiking Actor Network with Intra-layer Connections for
Reinforcement Learning [51.386945803485084]
We focus on the task where the agent needs to learn multi-dimensional deterministic policies to control.
Most existing spike-based RL methods take the firing rate as the output of SNNs, and convert it to represent continuous action space (i.e., the deterministic policy) through a fully-connected layer.
To develop a fully spiking actor network without any floating-point matrix operations, we draw inspiration from the non-spiking interneurons found in insects.
arXiv Detail & Related papers (2024-01-09T07:31:34Z) - Automating Continual Learning [42.710124929514066]
General-purpose learning systems should improve themselves in open-ended fashion in ever-changing environments.
We propose Automated Continual Learning (ACL) to train self-referential neural networks to meta-learn their own in-context continual (meta-)learning algorithms.
arXiv Detail & Related papers (2023-12-01T01:25:04Z) - MultIOD: Rehearsal-free Multihead Incremental Object Detector [17.236182938227163]
We propose MultIOD, a class-incremental object detector based on CenterNet.
We employ transfer learning between classes learned initially and those learned incrementally to tackle catastrophic forgetting.
Results show that our method outperforms state-of-the-art methods on two Pascal VOC datasets.
arXiv Detail & Related papers (2023-09-11T09:32:45Z) - Self-regulating Prompts: Foundational Model Adaptation without
Forgetting [112.66832145320434]
We introduce a self-regularization framework for prompting called PromptSRC.
PromptSRC guides the prompts to optimize for both task-specific and task-agnostic general representations.
arXiv Detail & Related papers (2023-07-13T17:59:35Z) - Artificial Neuronal Ensembles with Learned Context Dependent Gating [0.0]
We introduce Learned Context Dependent Gating (LXDG), a method to flexibly allocate and recall artificial neuronal ensembles'
Activities in the hidden layers of the network are modulated by gates, which are dynamically produced during training.
We demonstrate the ability of this method to alleviate catastrophic forgetting on continual learning benchmarks.
arXiv Detail & Related papers (2023-01-17T20:52:48Z) - Stacked unsupervised learning with a network architecture found by
supervised meta-learning [4.209801809583906]
Stacked unsupervised learning seems more biologically plausible than backpropagation.
But SUL has fallen far short of backpropagation in practical applications.
We show an SUL algorithm that can perform completely unsupervised clustering of MNIST digits.
arXiv Detail & Related papers (2022-06-06T16:17:20Z) - Learning Bayesian Sparse Networks with Full Experience Replay for
Continual Learning [54.7584721943286]
Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered.
Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal.
We propose to only activate and select sparse neurons for learning current and past tasks at any stage.
arXiv Detail & Related papers (2022-02-21T13:25:03Z) - Adaptive Rational Activations to Boost Deep Reinforcement Learning [68.10769262901003]
We motivate why rationals are suitable for adaptable activation functions and why their inclusion into neural networks is crucial.
We demonstrate that equipping popular algorithms with (recurrent-)rational activations leads to consistent improvements on Atari games.
arXiv Detail & Related papers (2021-02-18T14:53:12Z) - Towards Efficient Processing and Learning with Spikes: New Approaches
for Multi-Spike Learning [59.249322621035056]
We propose two new multi-spike learning rules which demonstrate better performance over other baselines on various tasks.
In the feature detection task, we re-examine the ability of unsupervised STDP with its limitations being presented.
Our proposed learning rules can reliably solve the task over a wide range of conditions without specific constraints being applied.
arXiv Detail & Related papers (2020-05-02T06:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.