SynthDST: Synthetic Data is All You Need for Few-Shot Dialog State
Tracking
- URL: http://arxiv.org/abs/2402.02285v1
- Date: Sat, 3 Feb 2024 22:49:00 GMT
- Title: SynthDST: Synthetic Data is All You Need for Few-Shot Dialog State
Tracking
- Authors: Atharva Kulkarni, Bo-Hsiang Tseng, Joel Ruben Antony Moniz, Dhivya
Piraviperumal, Hong Yu, Shruti Bhargava
- Abstract summary: In-context learning with Large Language Models (LLMs) has emerged as a promising avenue of research in Dialog State Tracking (DST)
We propose method, a data generation framework tailored for DST, utilizing LLMs.
Our approach only requires the dialogue schema and a few hand-crafted dialogue templates to synthesize natural, coherent, and free-flowing dialogues with DST annotations.
- Score: 12.300308969720916
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In-context learning with Large Language Models (LLMs) has emerged as a
promising avenue of research in Dialog State Tracking (DST). However, the
best-performing in-context learning methods involve retrieving and adding
similar examples to the prompt, requiring access to labeled training data.
Procuring such training data for a wide range of domains and applications is
time-consuming, expensive, and, at times, infeasible. While zero-shot learning
requires no training data, it significantly lags behind the few-shot setup.
Thus, `\textit{Can we efficiently generate synthetic data for any dialogue
schema to enable few-shot prompting?}' Addressing this question, we propose
\method, a data generation framework tailored for DST, utilizing LLMs. Our
approach only requires the dialogue schema and a few hand-crafted dialogue
templates to synthesize natural, coherent, and free-flowing dialogues with DST
annotations. Few-shot learning using data from {\method} results in $4-5%$
improvement in Joint Goal Accuracy over the zero-shot baseline on MultiWOZ 2.1
and 2.4. Remarkably, our few-shot learning approach recovers nearly $98%$ of
the performance compared to the few-shot setup using human-annotated training
data. Our synthetic data and code can be accessed at
https://github.com/apple/ml-synthdst
Related papers
- UNO-DST: Leveraging Unlabelled Data in Zero-Shot Dialogue State Tracking [54.51316566989655]
Previous zero-shot dialogue state tracking (DST) methods only apply transfer learning, ignoring unlabelled data in the target domain.
We transform zero-shot DST into few-shot DST by utilising such unlabelled data via joint and self-training methods.
We demonstrate this method's effectiveness on general language models in zero-shot scenarios, improving average joint goal accuracy by 8% across all domains in MultiWOZ.
arXiv Detail & Related papers (2023-10-16T15:16:16Z) - Diverse Retrieval-Augmented In-Context Learning for Dialogue State
Tracking [3.8073142980733]
We propose RefPyDST, which advances the state of the art with three advancements to in-context learning for dialogue state tracking.
First, we formulate DST as a Python programming task, explicitly modeling language coreference as variable reference in Python.
Second, since in-context learning depends highly on the context examples, we propose a method to retrieve a diverse set of relevant examples to improve performance.
arXiv Detail & Related papers (2023-07-04T03:15:52Z) - Stabilized In-Context Learning with Pre-trained Language Models for Few
Shot Dialogue State Tracking [57.92608483099916]
Large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks.
For more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial.
We introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query.
arXiv Detail & Related papers (2023-02-12T15:05:10Z) - Self-Prompting Large Language Models for Zero-Shot Open-Domain QA [67.08732962244301]
Open-Domain Question Answering (ODQA) aims to answer questions without explicitly providing background documents.
This task becomes notably challenging in a zero-shot setting where no data is available to train tailored retrieval-reader models.
We propose a Self-Prompting framework to explicitly utilize the massive knowledge encoded in the parameters of Large Language Models.
arXiv Detail & Related papers (2022-12-16T18:23:43Z) - Controllable Dialogue Simulation with In-Context Learning [39.04491297557292]
textscDialogic is a dialogue simulation method based on large language model in-context learning.
Our method can rapidly expand a small set of dialogue data with minimum or zero human involvement.
Our simulated dialogues have near-human fluency and annotation accuracy.
arXiv Detail & Related papers (2022-10-09T06:32:58Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Few-Shot Bot: Prompt-Based Learning for Dialogue Systems [58.27337673451943]
Learning to converse using only a few examples is a great challenge in conversational AI.
The current best conversational models are either good chit-chatters (e.g., BlenderBot) or goal-oriented systems (e.g., MinTL)
We propose prompt-based few-shot learning which does not require gradient-based fine-tuning but instead uses a few examples as the only source of learning.
arXiv Detail & Related papers (2021-10-15T14:36:45Z) - Language Models as Few-Shot Learner for Task-Oriented Dialogue Systems [74.8759568242933]
Task-oriented dialogue systems use four connected modules, namely, Natural Language Understanding (NLU), a Dialogue State Tracking (DST), Dialogue Policy (DP) and Natural Language Generation (NLG)
A research challenge is to learn each module with the least amount of samples given the high cost related to the data collection.
We evaluate the priming few-shot ability of language models in the NLU, DP and NLG tasks.
arXiv Detail & Related papers (2020-08-14T08:23:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.