Continual Prompt Tuning for Dialog State Tracking
- URL: http://arxiv.org/abs/2203.06654v1
- Date: Sun, 13 Mar 2022 13:22:41 GMT
- Title: Continual Prompt Tuning for Dialog State Tracking
- Authors: Qi Zhu, Bing Li, Fei Mi, Xiaoyan Zhu, Minlie Huang
- Abstract summary: A desirable dialog system should be able to continually learn new skills without forgetting old ones.
We present Continual Prompt Tuning, a parameter-efficient framework that not only avoids forgetting but also enables knowledge transfer between tasks.
- Score: 58.66412648276873
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A desirable dialog system should be able to continually learn new skills
without forgetting old ones, and thereby adapt to new domains or tasks in its
life cycle. However, continually training a model often leads to a well-known
catastrophic forgetting issue. In this paper, we present Continual Prompt
Tuning, a parameter-efficient framework that not only avoids forgetting but
also enables knowledge transfer between tasks. To avoid forgetting, we only
learn and store a few prompt tokens' embeddings for each task while freezing
the backbone pre-trained model. To achieve bi-directional knowledge transfer
among tasks, we propose several techniques (continual prompt initialization,
query fusion, and memory replay) to transfer knowledge from preceding tasks and
a memory-guided technique to transfer knowledge from subsequent tasks.
Extensive experiments demonstrate the effectiveness and efficiency of our
proposed method on continual learning for dialog state tracking, compared with
state-of-the-art baselines.
Related papers
- TaSL: Continual Dialog State Tracking via Task Skill Localization and Consolidation [14.533890076297393]
We present TaSL, a novel framework for task skill localization and consolidation.
TaSL uses a novel group-wise technique to pinpoint task-specific and task-shared areas.
As a result, TaSL strikes a balance between preserving previous knowledge and excelling at new tasks.
arXiv Detail & Related papers (2024-08-19T10:01:28Z) - PECTP: Parameter-Efficient Cross-Task Prompts for Incremental Vision Transformer [76.39111896665585]
Incremental Learning (IL) aims to learn deep models on sequential tasks continually.
Recent vast pre-trained models (PTMs) have achieved outstanding performance by prompt technique in practical IL without the old samples.
arXiv Detail & Related papers (2024-07-04T10:37:58Z) - Multitask Learning with No Regret: from Improved Confidence Bounds to
Active Learning [79.07658065326592]
Quantifying uncertainty in the estimated tasks is of pivotal importance for many downstream applications, such as online or active learning.
We provide novel multitask confidence intervals in the challenging setting when neither the similarity between tasks nor the tasks' features are available to the learner.
We propose a novel online learning algorithm that achieves such improved regret without knowing this parameter in advance.
arXiv Detail & Related papers (2023-08-03T13:08:09Z) - Online Continual Learning via the Knowledge Invariant and Spread-out
Properties [4.109784267309124]
Key challenge in continual learning is catastrophic forgetting.
We propose a new method, named Online Continual Learning via the Knowledge Invariant and Spread-out Properties (OCLKISP)
We empirically evaluate our proposed method on four popular benchmarks for continual learning: Split CIFAR 100, Split SVHN, Split CUB200 and Split Tiny-Image-Net.
arXiv Detail & Related papers (2023-02-02T04:03:38Z) - Beyond Not-Forgetting: Continual Learning with Backward Knowledge
Transfer [39.99577526417276]
In continual learning (CL) an agent can improve the learning performance of both a new task and old' tasks.
Most existing CL methods focus on addressing catastrophic forgetting in neural networks by minimizing the modification of the learnt model for old tasks.
We propose a new CL method with Backward knowlEdge tRansfer (CUBER) for a fixed capacity neural network without data replay.
arXiv Detail & Related papers (2022-11-01T23:55:51Z) - Relational Experience Replay: Continual Learning by Adaptively Tuning
Task-wise Relationship [54.73817402934303]
We propose Experience Continual Replay (ERR), a bi-level learning framework to adaptively tune task-wise to achieve a better stability plasticity' tradeoff.
ERR can consistently improve the performance of all baselines and surpass current state-of-the-art methods.
arXiv Detail & Related papers (2021-12-31T12:05:22Z) - Continual Learning in Task-Oriented Dialogue Systems [49.35627673523519]
Continual learning in task-oriented dialogue systems can allow us to add new domains and functionalities through time without incurring the high cost of a whole system retraining.
We propose a continual learning benchmark for task-oriented dialogue systems with 37 domains to be learned continuously in four settings.
arXiv Detail & Related papers (2020-12-31T08:44:25Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.