A Sequence-to-Sequence Approach to Dialogue State Tracking
- URL: http://arxiv.org/abs/2011.09553v2
- Date: Sat, 29 May 2021 02:48:28 GMT
- Title: A Sequence-to-Sequence Approach to Dialogue State Tracking
- Authors: Yue Feng, Yang Wang, Hang Li
- Abstract summary: Seq2Seq-DU formalizes dialogue state tracking as a sequence-to-sequence problem.
It can jointly model intents, slots, and slot values.
It can effectively deal with categorical and non-categorical slots, and unseen schemas.
- Score: 17.81139775400199
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper is concerned with dialogue state tracking (DST) in a task-oriented
dialogue system. Building a DST module that is highly effective is still a
challenging issue, although significant progresses have been made recently.
This paper proposes a new approach to dialogue state tracking, referred to as
Seq2Seq-DU, which formalizes DST as a sequence-to-sequence problem. Seq2Seq-DU
employs two BERT-based encoders to respectively encode the utterances in the
dialogue and the descriptions of schemas, an attender to calculate attentions
between the utterance embeddings and the schema embeddings, and a decoder to
generate pointers to represent the current state of dialogue. Seq2Seq-DU has
the following advantages. It can jointly model intents, slots, and slot values;
it can leverage the rich representations of utterances and schemas based on
BERT; it can effectively deal with categorical and non-categorical slots, and
unseen schemas. In addition, Seq2Seq-DU can also be used in the NLU (natural
language understanding) module of a dialogue system. Experimental results on
benchmark datasets in different settings (SGD, MultiWOZ2.2, MultiWOZ2.1,
WOZ2.0, DSTC2, M2M, SNIPS, and ATIS) show that Seq2Seq-DU outperforms the
existing methods.
Related papers
- Transforming Slot Schema Induction with Generative Dialogue State Inference [14.06505399101404]
Slot Induction (SSI) aims to automatically induce slots from unlabeled dialogue data.
Our SSI method discovers high-quality candidate information for representing dialogue state.
Experimental comparisons on the MultiWOZ and SGD datasets demonstrate that Generative Dialogue State Inference (GenDSI) outperforms the previous state-of-the-art.
arXiv Detail & Related papers (2024-08-03T02:41:10Z) - Diable: Efficient Dialogue State Tracking as Operations on Tables [12.750160147987186]
We propose a new task formalisation that simplifies the design and implementation of efficient dialogue state tracking systems.
We represent the dialogue state as a table and formalise DST as a table manipulation task.
At each turn, the system updates the previous state by generating table operations based on the dialogue context.
arXiv Detail & Related papers (2023-05-26T15:26:12Z) - A Multi-Task BERT Model for Schema-Guided Dialogue State Tracking [78.2700757742992]
Task-oriented dialogue systems often employ a Dialogue State Tracker (DST) to successfully complete conversations.
Recent state-of-the-art DST implementations rely on schemata of diverse services to improve model robustness.
We propose a single multi-task BERT-based model that jointly solves the three DST tasks of intent prediction, requested slot prediction and slot filling.
arXiv Detail & Related papers (2022-07-02T13:27:59Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Extended Graph Temporal Classification for Multi-Speaker End-to-End ASR [77.82653227783447]
We propose an extension of GTC to model the posteriors of both labels and label transitions by a neural network.
As an example application, we use the extended GTC (GTC-e) for the multi-speaker speech recognition task.
arXiv Detail & Related papers (2022-03-01T05:02:02Z) - Improving Mandarin End-to-End Speech Recognition with Word N-gram
Language Model [57.92200214957124]
External language models (LMs) are used to improve the recognition performance of end-to-end (E2E) automatic speech recognition (ASR) systems.
We propose a novel decoding algorithm where a word-level lattice is constructed on-the-fly to consider all possible word sequences.
Our method consistently outperforms subword-level LMs, including N-gram LM and neural network LM.
arXiv Detail & Related papers (2022-01-06T10:04:56Z) - Dialogue State Tracking with a Language Model using Schema-Driven
Prompting [18.83983018421701]
We introduce a new variation of the language modeling approach that uses schema-driven prompting to provide task-aware history encoding.
Our purely generative system achieves state-of-the-art performance on MultiWOZ 2.2 and achieves competitive performance on two other benchmarks: MultiWOZ 2.1 and M2M.
arXiv Detail & Related papers (2021-09-15T18:11:25Z) - Dual Learning for Dialogue State Tracking [44.679185483585364]
Dialogue state tracking (DST) is to estimate the dialogue state at each turn.
Due to the dependency on complicated dialogue history contexts, DST data annotation is more expensive than single-sentence language understanding.
We propose a novel dual-learning framework to make full use of unlabeled data.
arXiv Detail & Related papers (2020-09-22T10:15:09Z) - Video-Grounded Dialogues with Pretrained Generation Language Models [88.15419265622748]
We leverage the power of pre-trained language models for improving video-grounded dialogue.
We propose a framework by formulating sequence-to-grounded dialogue tasks as a sequence-to-grounded task.
Our framework allows fine-tuning language models to capture dependencies across multiple modalities.
arXiv Detail & Related papers (2020-06-27T08:24:26Z) - Efficient Context and Schema Fusion Networks for Multi-Domain Dialogue
State Tracking [32.36259992245]
Dialogue state tracking (DST) aims at estimating the current dialogue state given all the preceding conversation.
For multi-domain DST, the data sparsity problem is a major obstacle due to increased numbers of state candidates and dialogue lengths.
We utilize the previous dialogue state (predicted) and the current dialogue utterance as the input for DST.
arXiv Detail & Related papers (2020-04-07T13:46:39Z) - Non-Autoregressive Dialog State Tracking [122.2328875457225]
We propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST)
NADST can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots.
Our results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus.
arXiv Detail & Related papers (2020-02-19T06:39:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.