Prompt Learning for Few-Shot Dialogue State Tracking
- URL: http://arxiv.org/abs/2201.05780v1
- Date: Sat, 15 Jan 2022 07:37:33 GMT
- Title: Prompt Learning for Few-Shot Dialogue State Tracking
- Authors: Yuting Yang, Wenqiang Lei, Juan Cao, Jintao Li and Tat-Seng Chua
- Abstract summary: This paper focuses on how to learn a dialogue state tracking (DST) model efficiently with limited labeled data.
We design a prompt learning framework for few-shot DST, which consists of two main components: value-based prompt and inverse prompt mechanism.
Experiments show that our model can generate unseen slots and outperforms existing state-of-the-art few-shot methods.
- Score: 75.50701890035154
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Collecting dialogue state labels, slots and values, for learning dialogue
state tracking (DST) models can be costly, especially with the wide application
of dialogue systems in new-rising domains. In this paper, we focus on how to
learn a DST model efficiently with limited labeled data. We design a prompt
learning framework for few-shot DST, which consists of two main components:
value-based prompt and inverse prompt mechanism. This framework aims to utilize
the language understanding and generation ability of pre-trained language
models (PLM). First, we design value-based prompt functions to probe the
DST-related knowledge from PLM, which do not rely on the known ontology of
slots. Further, an inverse prompt mechanism is utilized to self-check the
"prompted" knowledge and help the PLM understand the essence of DST task
further. Experiments show that our model can generate unseen slots and
outperforms existing state-of-the-art few-shot methods. It indicates that
DST-related knowledge can be probed from PLM and utilized to address
low-resource DST efficiently with the help of prompt learning.
Related papers
- Training Zero-Shot Generalizable End-to-End Task-Oriented Dialog System Without Turn-level Dialog Annotations [2.757798192967912]
This work employs multi-task instruction fine-tuning to create more efficient and scalable task-oriented dialogue systems.
Our approach outperforms both state-of-the-art models trained on annotated data and billion-scale parameter off-the-shelf ChatGPT models.
arXiv Detail & Related papers (2024-07-21T04:52:38Z) - Injecting linguistic knowledge into BERT for Dialogue State Tracking [60.42231674887294]
This paper proposes a method that extracts linguistic knowledge via an unsupervised framework.
We then utilize this knowledge to augment BERT's performance and interpretability in Dialogue State Tracking (DST) tasks.
We benchmark this framework on various DST tasks and observe a notable improvement in accuracy.
arXiv Detail & Related papers (2023-11-27T08:38:42Z) - PromptAttack: Probing Dialogue State Trackers with Adversarial Prompts [25.467840115593784]
A key component of modern conversational systems is the Dialogue State Tracker (or DST)
We introduce a prompt-based learning approach to automatically generate effective adversarial examples to probe DST models.
We show how much the generated adversarial examples can bolster a DST through adversarial training.
arXiv Detail & Related papers (2023-06-07T15:41:40Z) - KILDST: Effective Knowledge-Integrated Learning for Dialogue State
Tracking using Gazetteer and Speaker Information [3.342637296393915]
Dialogue State Tracking (DST) is core research in dialogue systems and has received much attention.
It is necessary to define a new problem that can deal with dialogue between users as a step toward the conversational AI that extracts and recommends information from the dialogue between users.
We introduce a new task - DST from dialogue between users about scheduling an event (DST-S)
The DST-S task is much more challenging since it requires the model to understand and track dialogue in the dialogue between users and to understand who suggested the schedule and who agreed to the proposed schedule.
arXiv Detail & Related papers (2023-01-18T07:11:56Z) - A Study on Prompt-based Few-Shot Learning Methods for Belief State
Tracking in Task-oriented Dialog Systems [10.024834304960846]
We tackle the Dialogue Belief State Tracking problem of task-oriented conversational systems.
Recent approaches to this problem leveraging Transformer-based models have yielded great results.
We explore prompt-based few-shot learning for Dialogue Belief State Tracking.
arXiv Detail & Related papers (2022-04-18T05:29:54Z) - Knowledgeable Salient Span Mask for Enhancing Language Models as
Knowledge Base [51.55027623439027]
We develop two solutions to help the model learn more knowledge from unstructured text in a fully self-supervised manner.
To our best knowledge, we are the first to explore fully self-supervised learning of knowledge in continual pre-training.
arXiv Detail & Related papers (2022-04-17T12:33:34Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Improving Limited Labeled Dialogue State Tracking with Self-Supervision [91.68515201803986]
Existing dialogue state tracking (DST) models require plenty of labeled data.
We present and investigate two self-supervised objectives: preserving latent consistency and modeling conversational behavior.
Our proposed self-supervised signals can improve joint goal accuracy by 8.95% when only 1% labeled data is used.
arXiv Detail & Related papers (2020-10-26T21:57:42Z) - Language Models as Few-Shot Learner for Task-Oriented Dialogue Systems [74.8759568242933]
Task-oriented dialogue systems use four connected modules, namely, Natural Language Understanding (NLU), a Dialogue State Tracking (DST), Dialogue Policy (DP) and Natural Language Generation (NLG)
A research challenge is to learn each module with the least amount of samples given the high cost related to the data collection.
We evaluate the priming few-shot ability of language models in the NLU, DP and NLG tasks.
arXiv Detail & Related papers (2020-08-14T08:23:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.