TripPy: A Triple Copy Strategy for Value Independent Neural Dialog State
Tracking
- URL: http://arxiv.org/abs/2005.02877v4
- Date: Fri, 25 Sep 2020 13:46:29 GMT
- Title: TripPy: A Triple Copy Strategy for Value Independent Neural Dialog State
Tracking
- Authors: Michael Heck, Carel van Niekerk, Nurul Lubis, Christian Geishauser,
Hsien-Chin Lin, Marco Moresi, Milica Ga\v{s}i\'c
- Abstract summary: Task-oriented dialog systems rely on dialog state tracking (DST) to monitor the user's goal during an interaction.
We present a new approach to DST which makes use of various copy mechanisms to fill slots with values.
- Score: 2.78632567955797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Task-oriented dialog systems rely on dialog state tracking (DST) to monitor
the user's goal during the course of an interaction. Multi-domain and
open-vocabulary settings complicate the task considerably and demand scalable
solutions. In this paper we present a new approach to DST which makes use of
various copy mechanisms to fill slots with values. Our model has no need to
maintain a list of candidate values. Instead, all values are extracted from the
dialog context on-the-fly. A slot is filled by one of three copy mechanisms:
(1) Span prediction may extract values directly from the user input; (2) a
value may be copied from a system inform memory that keeps track of the
system's inform operations; (3) a value may be copied over from a different
slot that is already contained in the dialog state to resolve coreferences
within and across domains. Our approach combines the advantages of span-based
slot filling methods with memory methods to avoid the use of value picklists
altogether. We argue that our strategy simplifies the DST task while at the
same time achieving state of the art performance on various popular evaluation
sets including Multiwoz 2.1, where we achieve a joint goal accuracy beyond 55%.
Related papers
- DialCLIP: Empowering CLIP as Multi-Modal Dialog Retriever [83.33209603041013]
We propose a parameter-efficient prompt-tuning method named DialCLIP for multi-modal dialog retrieval.
Our approach introduces a multi-modal context generator to learn context features which are distilled into prompts within the pre-trained vision-language model CLIP.
To facilitate various types of retrieval, we also design multiple experts to learn mappings from CLIP outputs to multi-modal representation space.
arXiv Detail & Related papers (2024-01-02T07:40:12Z) - CGoDial: A Large-Scale Benchmark for Chinese Goal-oriented Dialog
Evaluation [75.60156479374416]
CGoDial is a new challenging and comprehensive Chinese benchmark for Goal-oriented Dialog evaluation.
It contains 96,763 dialog sessions and 574,949 dialog turns totally, covering three datasets with different knowledge sources.
To bridge the gap between academic benchmarks and spoken dialog scenarios, we either collect data from real conversations or add spoken features to existing datasets via crowd-sourcing.
arXiv Detail & Related papers (2022-11-21T16:21:41Z) - SPACE-3: Unified Dialog Model Pre-training for Task-Oriented Dialog
Understanding and Generation [123.37377363355363]
SPACE-3 is a novel unified semi-supervised pre-trained conversation model learning from large-scale dialog corpora.
It can be effectively fine-tuned on a wide range of downstream dialog tasks.
Results show that SPACE-3 achieves state-of-the-art performance on eight downstream dialog benchmarks.
arXiv Detail & Related papers (2022-09-14T14:17:57Z) - SPACE-2: Tree-Structured Semi-Supervised Contrastive Pre-training for
Task-Oriented Dialog Understanding [68.94808536012371]
We propose a tree-structured pre-trained conversation model, which learns dialog representations from limited labeled dialogs and large-scale unlabeled dialog corpora.
Our method can achieve new state-of-the-art results on the DialoGLUE benchmark consisting of seven datasets and four popular dialog understanding tasks.
arXiv Detail & Related papers (2022-09-14T13:42:50Z) - Ontology-Enhanced Slot Filling [3.024095356755364]
In multi-domain task-oriented dialog system, user utterances and system responses may mention multiple named entities and attributes values.
A system needs to select those that are confirmed by the user and fill them into destined slots.
One difficulty is that since a dialogue session contains multiple system-user turns, feeding in all the tokens into a deep model such as BERT can be challenging due to limited capacity of input word tokens and GPU memory.
arXiv Detail & Related papers (2021-08-25T14:54:47Z) - Zero-shot Generalization in Dialog State Tracking through Generative
Question Answering [10.81203437307028]
We introduce a novel framework that supports natural language queries for unseen constraints and slots in task-oriented dialogs.
Our approach is based on generative question-answering using a conditional domain model pre-trained on substantive English sentences.
arXiv Detail & Related papers (2021-01-20T21:47:20Z) - Dialog Simulation with Realistic Variations for Training Goal-Oriented
Conversational Systems [14.206866126142002]
Goal-oriented dialog systems enable users to complete specific goals like requesting information about a movie or booking a ticket.
We propose an approach for automatically creating a large corpus of annotated dialogs from a few thoroughly annotated sample dialogs and the dialog schema.
We achieve 18? 50% relative accuracy on a held-out test set compared to a baseline dialog generation approach.
arXiv Detail & Related papers (2020-11-16T19:39:15Z) - CREDIT: Coarse-to-Fine Sequence Generation for Dialogue State Tracking [44.38388988238695]
A dialogue state tracker aims to accurately find a compact representation of the current dialogue status.
We employ a structured state representation and cast dialogue state tracking as a sequence generation problem.
Experiments demonstrate our tracker achieves encouraging joint goal accuracy for the five domains in MultiWOZ 2.0 and MultiWOZ 2.1 datasets.
arXiv Detail & Related papers (2020-09-22T10:27:18Z) - DCR-Net: A Deep Co-Interactive Relation Network for Joint Dialog Act
Recognition and Sentiment Classification [77.59549450705384]
In dialog system, dialog act recognition and sentiment classification are two correlative tasks.
Most of the existing systems either treat them as separate tasks or just jointly model the two tasks.
We propose a Deep Co-Interactive Relation Network (DCR-Net) to explicitly consider the cross-impact and model the interaction between the two tasks.
arXiv Detail & Related papers (2020-08-16T14:13:32Z) - Non-Autoregressive Dialog State Tracking [122.2328875457225]
We propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST)
NADST can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots.
Our results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus.
arXiv Detail & Related papers (2020-02-19T06:39:26Z) - MA-DST: Multi-Attention Based Scalable Dialog State Tracking [13.358314140896937]
Dialog State Tracking dialog agents provide a natural language interface for users to complete their goal.
To enable accurate multi-domain DST, the model needs to encode dependencies between past utterances and slot semantics.
We introduce a novel architecture for this task to encode the conversation history and slot semantics.
arXiv Detail & Related papers (2020-02-07T05:34:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.