Dual Learning for Dialogue State Tracking
- URL: http://arxiv.org/abs/2009.10430v1
- Date: Tue, 22 Sep 2020 10:15:09 GMT
- Title: Dual Learning for Dialogue State Tracking
- Authors: Zhi Chen, Lu Chen, Yanbin Zhao, Su Zhu and Kai Yu
- Abstract summary: Dialogue state tracking (DST) is to estimate the dialogue state at each turn.
Due to the dependency on complicated dialogue history contexts, DST data annotation is more expensive than single-sentence language understanding.
We propose a novel dual-learning framework to make full use of unlabeled data.
- Score: 44.679185483585364
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In task-oriented multi-turn dialogue systems, dialogue state refers to a
compact representation of the user goal in the context of dialogue history.
Dialogue state tracking (DST) is to estimate the dialogue state at each turn.
Due to the dependency on complicated dialogue history contexts, DST data
annotation is more expensive than single-sentence language understanding, which
makes the task more challenging. In this work, we formulate DST as a sequence
generation problem and propose a novel dual-learning framework to make full use
of unlabeled data. In the dual-learning framework, there are two agents: the
primal tracker agent (utterance-to-state generator) and the dual utterance
generator agent (state-to-utterance genera-tor). Compared with traditional
supervised learning framework, dual learning can iteratively update both agents
through the reconstruction error and reward signal respectively without labeled
data. Reward sparsity problem is hard to solve in previous DST methods. In this
work, the reformulation of DST as a sequence generation model effectively
alleviates this problem. We call this primal tracker agent dual-DST.
Experimental results on MultiWOZ2.1 dataset show that the proposed dual-DST
works very well, especially when labelled data is limited. It achieves
comparable performance to the system where labeled data is fully used.
Related papers
- Enhancing Dialogue State Tracking Models through LLM-backed User-Agents Simulation [12.93942316816741]
GPT-4 is used to simulate the user and agent interaction, generating thousands of annotated dialogues with DST labels.
A two-stage fine-tuning on LLaMA 2 is performed on the generated data and the real data for the DST prediction.
Our approach is also capable of adapting to the dynamic demands in real-world scenarios, generating dialogues in new domains swiftly.
arXiv Detail & Related papers (2024-05-17T07:00:05Z) - CSS: Combining Self-training and Self-supervised Learning for Few-shot
Dialogue State Tracking [36.18207750352937]
Few-shot dialogue state tracking (DST) is a realistic problem that trains the DST model with limited labeled data.
We propose a few-shot DST framework called CSS, which Combines Self-training and Self-supervised learning methods.
Experimental results on the MultiWOZ dataset show that our proposed CSS achieves competitive performance in several few-shot scenarios.
arXiv Detail & Related papers (2022-10-11T04:55:16Z) - OPAL: Ontology-Aware Pretrained Language Model for End-to-End
Task-Oriented Dialogue [40.62090743056549]
This paper presents an ontology-aware pretrained language model (OPAL) for end-to-end task-oriented dialogue (TOD)
Unlike chit-chat dialogue models, task-oriented dialogue models fulfill at least two task-specific modules: dialogue state tracker (DST) and response generator (RG)
arXiv Detail & Related papers (2022-09-10T04:38:27Z) - A Model-Agnostic Data Manipulation Method for Persona-based Dialogue
Generation [107.82729587882397]
It is expensive to scale up current persona-based dialogue datasets.
Each data sample in this task is more complex to learn with than conventional dialogue data.
We propose a data manipulation method, which is model-agnostic to be packed with any persona-based dialogue generation model.
arXiv Detail & Related papers (2022-04-21T03:49:54Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Prompt Learning for Few-Shot Dialogue State Tracking [75.50701890035154]
This paper focuses on how to learn a dialogue state tracking (DST) model efficiently with limited labeled data.
We design a prompt learning framework for few-shot DST, which consists of two main components: value-based prompt and inverse prompt mechanism.
Experiments show that our model can generate unseen slots and outperforms existing state-of-the-art few-shot methods.
arXiv Detail & Related papers (2022-01-15T07:37:33Z) - Improving Longer-range Dialogue State Tracking [22.606650177804966]
Dialogue state tracking (DST) is a pivotal component in task-oriented dialogue systems.
In this paper, we aim to improve the overall performance of DST with a special focus on handling longer dialogues.
arXiv Detail & Related papers (2021-02-27T02:44:28Z) - Conditioned Text Generation with Transfer for Closed-Domain Dialogue
Systems [65.48663492703557]
We show how to optimally train and control the generation of intent-specific sentences using a conditional variational autoencoder.
We introduce a new protocol called query transfer that allows to leverage a large unlabelled dataset.
arXiv Detail & Related papers (2020-11-03T14:06:10Z) - Improving Limited Labeled Dialogue State Tracking with Self-Supervision [91.68515201803986]
Existing dialogue state tracking (DST) models require plenty of labeled data.
We present and investigate two self-supervised objectives: preserving latent consistency and modeling conversational behavior.
Our proposed self-supervised signals can improve joint goal accuracy by 8.95% when only 1% labeled data is used.
arXiv Detail & Related papers (2020-10-26T21:57:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.