Diable: Efficient Dialogue State Tracking as Operations on Tables
- URL: http://arxiv.org/abs/2305.17020v3
- Date: Wed, 1 Nov 2023 20:36:38 GMT
- Title: Diable: Efficient Dialogue State Tracking as Operations on Tables
- Authors: Pietro Lesci, Yoshinari Fujinuma, Momchil Hardalov, Chao Shang,
Yassine Benajiba, Lluis Marquez
- Abstract summary: We propose a new task formalisation that simplifies the design and implementation of efficient dialogue state tracking systems.
We represent the dialogue state as a table and formalise DST as a table manipulation task.
At each turn, the system updates the previous state by generating table operations based on the dialogue context.
- Score: 12.750160147987186
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Sequence-to-sequence state-of-the-art systems for dialogue state tracking
(DST) use the full dialogue history as input, represent the current state as a
list with all the slots, and generate the entire state from scratch at each
dialogue turn. This approach is inefficient, especially when the number of
slots is large and the conversation is long. We propose Diable, a new task
formalisation that simplifies the design and implementation of efficient DST
systems and allows one to easily plug and play large language models. We
represent the dialogue state as a table and formalise DST as a table
manipulation task. At each turn, the system updates the previous state by
generating table operations based on the dialogue context. Extensive
experimentation on the MultiWoz datasets demonstrates that Diable (i)
outperforms strong efficient DST baselines, (ii) is 2.4x more time efficient
than current state-of-the-art methods while retaining competitive Joint Goal
Accuracy, and (iii) is robust to noisy data annotations due to the table
operations approach.
Related papers
- TransferTOD: A Generalizable Chinese Multi-Domain Task-Oriented Dialogue System with Transfer Capabilities [46.91749457402889]
Task-oriented dialogue (TOD) systems aim to efficiently handle task-oriented conversations, including information collection.
How to utilize TOD accurately, efficiently and effectively for information collection has always been a critical and challenging task.
Recent studies have demonstrated that Large Language Models (LLMs) excel in dialogue, instruction generation, and reasoning.
arXiv Detail & Related papers (2024-07-31T15:38:15Z) - TOD-Flow: Modeling the Structure of Task-Oriented Dialogues [77.15457469745364]
We propose a novel approach focusing on inferring the TOD-Flow graph from dialogue data annotated with dialog acts.
The inferred TOD-Flow graph can be easily integrated with any dialogue model to improve its prediction performance, transparency, and controllability.
arXiv Detail & Related papers (2023-12-07T20:06:23Z) - OPAL: Ontology-Aware Pretrained Language Model for End-to-End
Task-Oriented Dialogue [40.62090743056549]
This paper presents an ontology-aware pretrained language model (OPAL) for end-to-end task-oriented dialogue (TOD)
Unlike chit-chat dialogue models, task-oriented dialogue models fulfill at least two task-specific modules: dialogue state tracker (DST) and response generator (RG)
arXiv Detail & Related papers (2022-09-10T04:38:27Z) - A Multi-Task BERT Model for Schema-Guided Dialogue State Tracking [78.2700757742992]
Task-oriented dialogue systems often employ a Dialogue State Tracker (DST) to successfully complete conversations.
Recent state-of-the-art DST implementations rely on schemata of diverse services to improve model robustness.
We propose a single multi-task BERT-based model that jointly solves the three DST tasks of intent prediction, requested slot prediction and slot filling.
arXiv Detail & Related papers (2022-07-02T13:27:59Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - A Sequence-to-Sequence Approach to Dialogue State Tracking [17.81139775400199]
Seq2Seq-DU formalizes dialogue state tracking as a sequence-to-sequence problem.
It can jointly model intents, slots, and slot values.
It can effectively deal with categorical and non-categorical slots, and unseen schemas.
arXiv Detail & Related papers (2020-11-18T21:42:44Z) - Task-Oriented Dialogue as Dataflow Synthesis [158.77123205487334]
We describe an approach to task-oriented dialogue in which dialogue state is represented as a dataflow graph.
A dialogue agent maps each user utterance to a program that extends this graph.
We introduce a new dataset, SMCalFlow, featuring complex dialogues about events, weather, places, and people.
arXiv Detail & Related papers (2020-09-24T00:35:26Z) - CREDIT: Coarse-to-Fine Sequence Generation for Dialogue State Tracking [44.38388988238695]
A dialogue state tracker aims to accurately find a compact representation of the current dialogue status.
We employ a structured state representation and cast dialogue state tracking as a sequence generation problem.
Experiments demonstrate our tracker achieves encouraging joint goal accuracy for the five domains in MultiWOZ 2.0 and MultiWOZ 2.1 datasets.
arXiv Detail & Related papers (2020-09-22T10:27:18Z) - Dual Learning for Dialogue State Tracking [44.679185483585364]
Dialogue state tracking (DST) is to estimate the dialogue state at each turn.
Due to the dependency on complicated dialogue history contexts, DST data annotation is more expensive than single-sentence language understanding.
We propose a novel dual-learning framework to make full use of unlabeled data.
arXiv Detail & Related papers (2020-09-22T10:15:09Z) - Non-Autoregressive Dialog State Tracking [122.2328875457225]
We propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST)
NADST can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots.
Our results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus.
arXiv Detail & Related papers (2020-02-19T06:39:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.