Turn-Level Active Learning for Dialogue State Tracking
- URL: http://arxiv.org/abs/2310.14513v1
- Date: Mon, 23 Oct 2023 02:53:46 GMT
- Title: Turn-Level Active Learning for Dialogue State Tracking
- Authors: Zihan Zhang, Meng Fang, Fanghua Ye, Ling Chen, Mohammad-Reza
Namazi-Rad
- Abstract summary: Dialogue state tracking (DST) plays an important role in task-oriented dialogue systems.
We propose a novel turn-level active learning framework for DST to actively select turns in dialogues to annotate.
- Score: 44.752369492979064
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Dialogue state tracking (DST) plays an important role in task-oriented
dialogue systems. However, collecting a large amount of turn-by-turn annotated
dialogue data is costly and inefficient. In this paper, we propose a novel
turn-level active learning framework for DST to actively select turns in
dialogues to annotate. Given the limited labelling budget, experimental results
demonstrate the effectiveness of selective annotation of dialogue turns.
Additionally, our approach can effectively achieve comparable DST performance
to traditional training approaches with significantly less annotated data,
which provides a more efficient way to annotate new dialogue data.
Related papers
- Data Augmentation Integrating Dialogue Flow and Style to Adapt Spoken Dialogue Systems to Low-Resource User Groups [1.7725414095035827]
This study addresses the interaction challenges encountered by spoken dialogue systems (SDSs) when engaging with users who exhibit distinct conversational behaviors.
We propose a novel data augmentation framework to enhance SDS performance for user groups with limited resources.
arXiv Detail & Related papers (2024-08-20T03:33:04Z) - Self-Explanation Prompting Improves Dialogue Understanding in Large
Language Models [52.24756457516834]
We propose a novel "Self-Explanation" prompting strategy to enhance the comprehension abilities of Large Language Models (LLMs)
This task-agnostic approach requires the model to analyze each dialogue utterance before task execution, thereby improving performance across various dialogue-centric tasks.
Experimental results from six benchmark datasets confirm that our method consistently outperforms other zero-shot prompts and matches or exceeds the efficacy of few-shot prompts.
arXiv Detail & Related papers (2023-09-22T15:41:34Z) - Continual Dialogue State Tracking via Example-Guided Question Answering [48.31523413835549]
We propose reformulating dialogue state tracking as a bundle of granular example-guided question answering tasks.
Our approach alleviates service-specific memorization and teaches a model to contextualize the given question and example.
We find that a model with just 60M parameters can achieve a significant boost by learning to learn from in-context examples retrieved by a retriever trained to identify turns with similar dialogue state changes.
arXiv Detail & Related papers (2023-05-23T06:15:43Z) - Few-Shot Dialogue Summarization via Skeleton-Assisted Prompt Transfer in
Prompt Tuning [47.336815771549524]
Skeleton-Assisted Prompt Transfer improves prompt transfer from dialogue state tracking to dialogue summarization.
We propose a novel approach with perturbation-based probes requiring neither annotation effort nor domain knowledge.
In-depth analyses demonstrate the effectiveness of our method in facilitating cross-task knowledge transfer in few-shot dialogue summarization.
arXiv Detail & Related papers (2023-05-20T03:32:48Z) - Dialogue State Distillation Network with Inter-Slot Contrastive Learning
for Dialogue State Tracking [25.722458066685046]
Dialogue State Tracking (DST) aims to extract users' intentions from the dialogue history.
Currently, most existing approaches suffer from error propagation and are unable to dynamically select relevant information.
We propose a Dialogue State Distillation Network (DSDN) to utilize relevant information of previous dialogue states.
arXiv Detail & Related papers (2023-02-16T11:05:24Z) - KILDST: Effective Knowledge-Integrated Learning for Dialogue State
Tracking using Gazetteer and Speaker Information [3.342637296393915]
Dialogue State Tracking (DST) is core research in dialogue systems and has received much attention.
It is necessary to define a new problem that can deal with dialogue between users as a step toward the conversational AI that extracts and recommends information from the dialogue between users.
We introduce a new task - DST from dialogue between users about scheduling an event (DST-S)
The DST-S task is much more challenging since it requires the model to understand and track dialogue in the dialogue between users and to understand who suggested the schedule and who agreed to the proposed schedule.
arXiv Detail & Related papers (2023-01-18T07:11:56Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Structure Extraction in Task-Oriented Dialogues with Slot Clustering [94.27806592467537]
In task-oriented dialogues, dialogue structure has often been considered as transition graphs among dialogue states.
We propose a simple yet effective approach for structure extraction in task-oriented dialogues.
arXiv Detail & Related papers (2022-02-28T20:18:12Z) - Improving Longer-range Dialogue State Tracking [22.606650177804966]
Dialogue state tracking (DST) is a pivotal component in task-oriented dialogue systems.
In this paper, we aim to improve the overall performance of DST with a special focus on handling longer dialogues.
arXiv Detail & Related papers (2021-02-27T02:44:28Z) - Modeling Long Context for Task-Oriented Dialogue State Generation [51.044300192906995]
We propose a multi-task learning model with a simple yet effective utterance tagging technique and a bidirectional language model.
Our approaches attempt to solve the problem that the performance of the baseline significantly drops when the input dialogue context sequence is long.
In our experiments, our proposed model achieves a 7.03% relative improvement over the baseline, establishing a new state-of-the-art joint goal accuracy of 52.04% on the MultiWOZ 2.0 dataset.
arXiv Detail & Related papers (2020-04-29T11:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.