Few-shot Continual Learning: a Brain-inspired Approach
- URL: http://arxiv.org/abs/2104.09034v1
- Date: Mon, 19 Apr 2021 03:40:48 GMT
- Title: Few-shot Continual Learning: a Brain-inspired Approach
- Authors: Liyuan Wang, Qian Li, Yi Zhong and Jun Zhu
- Abstract summary: We provide a first systematic study on few-shot continual learning (FSCL) and present an effective solution with deep neural networks.
Our solution is based on the observation that continual learning of a task sequence inevitably interferes few-shot generalization.
We draw inspirations from the robust brain system and develop a method that (1) interdependently updates a pair of fast / slow weights for continual learning and few-shot learning to disentangle their divergent objectives, inspired by the biological model of meta-plasticity and fast / slow synapse; and (2) applies a brain-inspired two-step consolidation strategy to learn a task sequence without forgetting in the
- Score: 34.306678703379944
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is an important yet challenging setting to continually learn new tasks
from a few examples. Although numerous efforts have been devoted to either
continual learning or few-shot learning, little work has considered this new
setting of few-shot continual learning (FSCL), which needs to minimize the
catastrophic forgetting to the old tasks and gradually improve the ability of
few-shot generalization. In this paper, we provide a first systematic study on
FSCL and present an effective solution with deep neural networks. Our solution
is based on the observation that continual learning of a task sequence
inevitably interferes few-shot generalization, which makes it highly nontrivial
to extend few-shot learning strategies to continual learning scenarios. We draw
inspirations from the robust brain system and develop a method that (1)
interdependently updates a pair of fast / slow weights for continual learning
and few-shot learning to disentangle their divergent objectives, inspired by
the biological model of meta-plasticity and fast / slow synapse; and (2)
applies a brain-inspired two-step consolidation strategy to learn a task
sequence without forgetting in the fast weights while improve generalization
without overfitting in the slow weights. Extensive results on various
benchmarks show that our method achieves a better performance than joint
training of all the tasks ever seen. The ability of few-shot generalization is
also substantially improved from incoming tasks and examples.
Related papers
- Simple and Effective Transfer Learning for Neuro-Symbolic Integration [50.592338727912946]
A potential solution to this issue is Neuro-Symbolic Integration (NeSy), where neural approaches are combined with symbolic reasoning.
Most of these methods exploit a neural network to map perceptions to symbols and a logical reasoner to predict the output of the downstream task.
They suffer from several issues, including slow convergence, learning difficulties with complex perception tasks, and convergence to local minima.
This paper proposes a simple yet effective method to ameliorate these problems.
arXiv Detail & Related papers (2024-02-21T15:51:01Z) - Continual Few-Shot Learning with Adversarial Class Storage [44.04528506999142]
We propose Continual Meta-Learner (CML) to solve the continual few-shot learning problem.
CML integrates metric-based classification and a memory-based mechanism along with adversarial learning into a meta-learning framework.
Experimental results show that CML delivers state-of-the-art performance on few-shot learning tasks without catastrophic forgetting.
arXiv Detail & Related papers (2022-07-10T03:40:38Z) - On the Effectiveness of Fine-tuning Versus Meta-reinforcement Learning [71.55412580325743]
We show that multi-task pretraining with fine-tuning on new tasks performs equally as well, or better, than meta-pretraining with meta test-time adaptation.
This is encouraging for future research, as multi-task pretraining tends to be simpler and computationally cheaper than meta-RL.
arXiv Detail & Related papers (2022-06-07T13:24:00Z) - CoMPS: Continual Meta Policy Search [113.33157585319906]
We develop a new continual meta-learning method to address challenges in sequential multi-task learning.
We find that CoMPS outperforms prior continual learning and off-policy meta-reinforcement methods on several sequences of challenging continuous control tasks.
arXiv Detail & Related papers (2021-12-08T18:53:08Z) - Towards More Generalizable One-shot Visual Imitation Learning [81.09074706236858]
A general-purpose robot should be able to master a wide range of tasks and quickly learn a novel one by leveraging past experiences.
One-shot imitation learning (OSIL) approaches this goal by training an agent with (pairs of) expert demonstrations.
We push for a higher level of generalization ability by investigating a more ambitious multi-task setup.
arXiv Detail & Related papers (2021-10-26T05:49:46Z) - AFEC: Active Forgetting of Negative Transfer in Continual Learning [37.03139674884091]
We show that biological neural networks can actively forget the old knowledge that conflicts with the learning of a new experience.
Inspired by the biological active forgetting, we propose to actively forget the old knowledge that limits the learning of new tasks to benefit continual learning.
arXiv Detail & Related papers (2021-10-23T10:03:19Z) - Lifelong Learning of Few-shot Learners across NLP Tasks [45.273018249235705]
We study the challenge of lifelong learning to few-shot learn over a sequence of diverse NLP tasks.
We propose a continual meta-learning approach which learns to generate adapter weights from a few examples.
We demonstrate our approach preserves model performance over training tasks and leads to positive knowledge transfer when the future tasks are learned.
arXiv Detail & Related papers (2021-04-18T10:41:56Z) - Importance Weighted Policy Learning and Adaptation [89.46467771037054]
We study a complementary approach which is conceptually simple, general, modular and built on top of recent improvements in off-policy learning.
The framework is inspired by ideas from the probabilistic inference literature and combines robust off-policy learning with a behavior prior.
Our approach achieves competitive adaptation performance on hold-out tasks compared to meta reinforcement learning baselines and can scale to complex sparse-reward scenarios.
arXiv Detail & Related papers (2020-09-10T14:16:58Z) - Understanding the Role of Training Regimes in Continual Learning [51.32945003239048]
Catastrophic forgetting affects the training of neural networks, limiting their ability to learn multiple tasks sequentially.
We study the effect of dropout, learning rate decay, and batch size, on forming training regimes that widen the tasks' local minima.
arXiv Detail & Related papers (2020-06-12T06:00:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.