How Relevant is Selective Memory Population in Lifelong Language
Learning?
- URL: http://arxiv.org/abs/2210.00940v1
- Date: Mon, 3 Oct 2022 13:52:54 GMT
- Title: How Relevant is Selective Memory Population in Lifelong Language
Learning?
- Authors: Vladimir Araujo, Helena Balabin, Julio Hurtado, Alvaro Soto,
Marie-Francine Moens
- Abstract summary: State-of-the-art approaches rely on sparse experience replay as the primary approach to prevent forgetting.
We investigate how relevant the selective memory population is in the lifelong learning process of text classification and question-answering tasks.
- Score: 15.9310767099639
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lifelong language learning seeks to have models continuously learn multiple
tasks in a sequential order without suffering from catastrophic forgetting.
State-of-the-art approaches rely on sparse experience replay as the primary
approach to prevent forgetting. Experience replay usually adopts sampling
methods for the memory population; however, the effect of the chosen sampling
strategy on model performance has not yet been studied. In this paper, we
investigate how relevant the selective memory population is in the lifelong
learning process of text classification and question-answering tasks. We found
that methods that randomly store a uniform number of samples from the entire
data stream lead to high performances, especially for low memory size, which is
consistent with computer vision studies.
Related papers
- Cache & Distil: Optimising API Calls to Large Language Models [82.32065572907125]
Large-scale deployment of generative AI tools often depends on costly API calls to a Large Language Model (LLM) to fulfil user queries.
To curtail the frequency of these calls, one can employ a smaller language model -- a student.
This student gradually gains proficiency in independently handling an increasing number of user requests.
arXiv Detail & Related papers (2023-10-20T15:01:55Z) - Improving Replay Sample Selection and Storage for Less Forgetting in
Continual Learning [1.2891210250935146]
Continual learning seeks to enable deep learners to train on a series of tasks of unknown length without suffering from the catastrophic forgetting of previous tasks.
One effective solution is replay, which involves storing few previous experiences in memory and replaying them when learning the current task.
This study aims to address these issues with a novel comparison of the commonly used reservoir sampling to various alternative population strategies.
arXiv Detail & Related papers (2023-08-03T17:46:27Z) - Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong
Learning in Task-Oriented Dialogue [80.05509768165135]
generative replay methods are widely employed to consolidate past knowledge with generated pseudo samples.
Most existing generative replay methods use only a single task-specific token to control their models.
We propose a novel method, prompt conditioned VAE for lifelong learning, to enhance generative replay by incorporating tasks' statistics.
arXiv Detail & Related papers (2022-10-14T13:12:14Z) - A Memory Transformer Network for Incremental Learning [64.0410375349852]
We study class-incremental learning, a training setup in which new classes of data are observed over time for the model to learn from.
Despite the straightforward problem formulation, the naive application of classification models to class-incremental learning results in the "catastrophic forgetting" of previously seen classes.
One of the most successful existing methods has been the use of a memory of exemplars, which overcomes the issue of catastrophic forgetting by saving a subset of past data into a memory bank and utilizing it to prevent forgetting when training future tasks.
arXiv Detail & Related papers (2022-10-10T08:27:28Z) - Memory Population in Continual Learning via Outlier Elimination [25.511380924335207]
Catastrophic forgetting, the phenomenon of forgetting previously learned tasks when learning a new one, is a major hurdle in developing continual learning algorithms.
A popular method to alleviate forgetting is to use a memory buffer, which stores a subset of previously learned task examples for use during training on new tasks.
This paper introduces Memory Outlier Elimination (MOE), a method for identifying and eliminating outliers in the memory buffer by choosing samples from label-homogeneous subpopulations.
arXiv Detail & Related papers (2022-07-04T00:09:33Z) - A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental
Learning [56.450090618578]
Class-Incremental Learning (CIL) aims to train a model with limited memory size to meet this requirement.
We show that when counting the model size into the total budget and comparing methods with aligned memory size, saving models do not consistently work.
We propose a simple yet effective baseline, denoted as MEMO for Memory-efficient Expandable MOdel.
arXiv Detail & Related papers (2022-05-26T08:24:01Z) - Memory Efficient Continual Learning for Neural Text Classification [10.70710638820641]
We devise a method to perform text classification using pre-trained models on a sequence of classification tasks provided in sequence.
We empirically demonstrate that our method requires significantly less model parameters compared to other state of the art methods.
While our method suffers little forgetting, it retains a predictive performance on-par with state of the art but less memory efficient methods.
arXiv Detail & Related papers (2022-03-09T10:57:59Z) - Towards Lifelong Learning of Multilingual Text-To-Speech Synthesis [87.75833205560406]
This work presents a lifelong learning approach to train a multilingual Text-To-Speech (TTS) system.
It does not require pooled data from all languages altogether, and thus alleviates the storage and computation burden.
arXiv Detail & Related papers (2021-10-09T07:00:38Z) - Efficient Meta Lifelong-Learning with Limited Memory [10.877225692975886]
State-of-the-art lifelong language learning methods store past examples in episodic memory and replay them at both training and inference time.
We identify three common principles of lifelong learning methods and propose an efficient meta-lifelong framework that combines them in a synergistic fashion.
arXiv Detail & Related papers (2020-10-06T06:08:07Z) - Meta-Learning with Sparse Experience Replay for Lifelong Language
Learning [26.296412053816233]
We propose a novel approach to lifelong learning of language tasks based on meta-learning with sparse experience replay.
We show that under the realistic setting of performing a single pass on a stream of tasks, our method obtains state-of-the-art results on lifelong text classification and relation extraction.
arXiv Detail & Related papers (2020-09-10T14:36:38Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.