Representation Memorization for Fast Learning New Knowledge without
Forgetting
- URL: http://arxiv.org/abs/2108.12596v1
- Date: Sat, 28 Aug 2021 07:54:53 GMT
- Title: Representation Memorization for Fast Learning New Knowledge without
Forgetting
- Authors: Fei Mi, Tao Lin, and Boi Faltings
- Abstract summary: The ability to quickly learn new knowledge is a big step towards human-level intelligence.
We consider scenarios that require learning new classes or data distributions quickly and incrementally over time.
We propose "Memory-based Hebbian Adaptation" to tackle the two major challenges.
- Score: 36.55736909586313
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability to quickly learn new knowledge (e.g. new classes or data
distributions) is a big step towards human-level intelligence. In this paper,
we consider scenarios that require learning new classes or data distributions
quickly and incrementally over time, as it often occurs in real-world dynamic
environments. We propose "Memory-based Hebbian Parameter Adaptation" (Hebb) to
tackle the two major challenges (i.e., catastrophic forgetting and sample
efficiency) towards this goal in a unified framework. To mitigate catastrophic
forgetting, Hebb augments a regular neural classifier with a continuously
updated memory module to store representations of previous data. To improve
sample efficiency, we propose a parameter adaptation method based on the
well-known Hebbian theory, which directly "wires" the output network's
parameters with similar representations retrieved from the memory. We
empirically verify the superior performance of Hebb through extensive
experiments on a wide range of learning tasks (image classification, language
model) and learning scenarios (continual, incremental, online). We demonstrate
that Hebb effectively mitigates catastrophic forgetting, and it indeed learns
new knowledge better and faster than the current state-of-the-art.
Related papers
- Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Complementary Learning Subnetworks for Parameter-Efficient
Class-Incremental Learning [40.13416912075668]
We propose a rehearsal-free CIL approach that learns continually via the synergy between two Complementary Learning Subnetworks.
Our method achieves competitive results against state-of-the-art methods, especially in accuracy gain, memory cost, training efficiency, and task-order.
arXiv Detail & Related papers (2023-06-21T01:43:25Z) - Error Sensitivity Modulation based Experience Replay: Mitigating Abrupt
Representation Drift in Continual Learning [13.041607703862724]
We propose ESMER, which employs a principled mechanism to modulate error sensitivity in a dual-memory rehearsal-based system.
ESMER effectively reduces forgetting and abrupt drift in representations at the task boundary by gradually adapting to the new task while consolidating knowledge.
Remarkably, it also enables the model to learn under high levels of label noise, which is ubiquitous in real-world data streams.
arXiv Detail & Related papers (2023-02-14T16:35:54Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Data augmentation on-the-fly and active learning in data stream
classification [9.367903089535684]
There is an emerging need for predictive models to be trained on-the-fly.
Learning models have access to more labelled data without the need to increase the active learning budget.
Augmented Queues significantly improves the performance in terms of learning quality and speed.
arXiv Detail & Related papers (2022-10-13T09:57:08Z) - A Memory Transformer Network for Incremental Learning [64.0410375349852]
We study class-incremental learning, a training setup in which new classes of data are observed over time for the model to learn from.
Despite the straightforward problem formulation, the naive application of classification models to class-incremental learning results in the "catastrophic forgetting" of previously seen classes.
One of the most successful existing methods has been the use of a memory of exemplars, which overcomes the issue of catastrophic forgetting by saving a subset of past data into a memory bank and utilizing it to prevent forgetting when training future tasks.
arXiv Detail & Related papers (2022-10-10T08:27:28Z) - New Insights on Reducing Abrupt Representation Change in Online
Continual Learning [69.05515249097208]
We focus on the change in representations of observed data that arises when previously unobserved classes appear in the incoming data stream.
We show that applying Experience Replay causes the newly added classes' representations to overlap significantly with the previous classes.
We propose a new method which mitigates this issue by shielding the learned representations from drastic adaptation to accommodate new classes.
arXiv Detail & Related papers (2022-03-08T01:37:00Z) - ZS-IL: Looking Back on Learned ExperiencesFor Zero-Shot Incremental
Learning [9.530976792843495]
We propose an on-call transfer set to provide past experiences whenever a new class arises in the data stream.
ZS-IL demonstrates significantly better performance on the well-known datasets (CIFAR-10, Tiny-ImageNet) in both Task-IL and Class-IL settings.
arXiv Detail & Related papers (2021-03-22T22:43:20Z) - Learning to Learn Variational Semantic Memory [132.39737669936125]
We introduce variational semantic memory into meta-learning to acquire long-term knowledge for few-shot learning.
The semantic memory is grown from scratch and gradually consolidated by absorbing information from tasks it experiences.
We formulate memory recall as the variational inference of a latent memory variable from addressed contents.
arXiv Detail & Related papers (2020-10-20T15:05:26Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.