Recall-Oriented Continual Learning with Generative Adversarial
Meta-Model
- URL: http://arxiv.org/abs/2403.03082v1
- Date: Tue, 5 Mar 2024 16:08:59 GMT
- Title: Recall-Oriented Continual Learning with Generative Adversarial
Meta-Model
- Authors: Haneol Kang, Dong-Wan Choi
- Abstract summary: We propose a recall-oriented continual learning framework to address the stability-plasticity dilemma.
Inspired by the human brain's ability to separate the mechanisms responsible for stability and plasticity, our framework consists of a two-level architecture.
We show that our framework not only effectively learns new knowledge without any disruption but also achieves high stability of previous knowledge.
- Score: 5.710971447109951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The stability-plasticity dilemma is a major challenge in continual learning,
as it involves balancing the conflicting objectives of maintaining performance
on previous tasks while learning new tasks. In this paper, we propose the
recall-oriented continual learning framework to address this challenge.
Inspired by the human brain's ability to separate the mechanisms responsible
for stability and plasticity, our framework consists of a two-level
architecture where an inference network effectively acquires new knowledge and
a generative network recalls past knowledge when necessary. In particular, to
maximize the stability of past knowledge, we investigate the complexity of
knowledge depending on different representations, and thereby introducing
generative adversarial meta-model (GAMM) that incrementally learns
task-specific parameters instead of input data samples of the task. Through our
experiments, we show that our framework not only effectively learns new
knowledge without any disruption but also achieves high stability of previous
knowledge in both task-aware and task-agnostic learning scenarios. Our code is
available at: https://github.com/bigdata-inha/recall-oriented-cl-framework.
Related papers
- A Unified Framework for Continual Learning and Machine Unlearning [9.538733681436836]
Continual learning and machine unlearning are crucial challenges in machine learning, typically addressed separately.
We introduce a novel framework that jointly tackles both tasks by leveraging controlled knowledge distillation.
Our approach enables efficient learning with minimal forgetting and effective targeted unlearning.
arXiv Detail & Related papers (2024-08-21T06:49:59Z) - Informed Meta-Learning [55.2480439325792]
Meta-learning and informed ML stand out as two approaches for incorporating prior knowledge into ML pipelines.
We formalise a hybrid paradigm, informed meta-learning, facilitating the incorporation of priors from unstructured knowledge representations.
We demonstrate the potential benefits of informed meta-learning in improving data efficiency, robustness to observational noise and task distribution shifts.
arXiv Detail & Related papers (2024-02-25T15:08:37Z) - Online Continual Learning via the Knowledge Invariant and Spread-out
Properties [4.109784267309124]
Key challenge in continual learning is catastrophic forgetting.
We propose a new method, named Online Continual Learning via the Knowledge Invariant and Spread-out Properties (OCLKISP)
We empirically evaluate our proposed method on four popular benchmarks for continual learning: Split CIFAR 100, Split SVHN, Split CUB200 and Split Tiny-Image-Net.
arXiv Detail & Related papers (2023-02-02T04:03:38Z) - Adaptively Integrated Knowledge Distillation and Prediction Uncertainty
for Continual Learning [71.43841235954453]
Current deep learning models often suffer from catastrophic forgetting of old knowledge when continually learning new knowledge.
Existing strategies to alleviate this issue often fix the trade-off between keeping old knowledge (stability) and learning new knowledge (plasticity)
arXiv Detail & Related papers (2023-01-18T05:36:06Z) - Progressive Learning without Forgetting [8.563323015260709]
We focus on two challenging problems in the paradigm of Continual Learning (CL)
PLwF introduces functions from previous tasks to construct a knowledge space that contains the most reliable knowledge on each task.
Credit assignment controls the tug-of-war dynamics by removing gradient conflict through projection.
In comparison with other CL methods, we report notably better results even without relying on any raw data.
arXiv Detail & Related papers (2022-11-28T10:53:14Z) - Anti-Retroactive Interference for Lifelong Learning [65.50683752919089]
We design a paradigm for lifelong learning based on meta-learning and associative mechanism of the brain.
It tackles the problem from two aspects: extracting knowledge and memorizing knowledge.
It is theoretically analyzed that the proposed learning paradigm can make the models of different tasks converge to the same optimum.
arXiv Detail & Related papers (2022-08-27T09:27:36Z) - Continual Prompt Tuning for Dialog State Tracking [58.66412648276873]
A desirable dialog system should be able to continually learn new skills without forgetting old ones.
We present Continual Prompt Tuning, a parameter-efficient framework that not only avoids forgetting but also enables knowledge transfer between tasks.
arXiv Detail & Related papers (2022-03-13T13:22:41Z) - Addressing the Stability-Plasticity Dilemma via Knowledge-Aware
Continual Learning [5.979373021392084]
We show that being aware of existing knowledge helps in: (1) increasing the forward transfer from similar knowledge, (2) reducing the required capacity by leveraging existing knowledge, and (4) increasing robustness to the class order in the sequence.
We evaluate sequences of similar tasks, dissimilar tasks, and a mix of both constructed from the two commonly used benchmarks for class-incremental learning; CIFAR-10 and CIFAR-100.
arXiv Detail & Related papers (2021-10-11T14:51:56Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z) - Automated Relational Meta-learning [95.02216511235191]
We propose an automated relational meta-learning framework that automatically extracts the cross-task relations and constructs the meta-knowledge graph.
We conduct extensive experiments on 2D toy regression and few-shot image classification and the results demonstrate the superiority of ARML over state-of-the-art baselines.
arXiv Detail & Related papers (2020-01-03T07:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.