Description-Driven Task-Oriented Dialog Modeling
- URL: http://arxiv.org/abs/2201.08904v1
- Date: Fri, 21 Jan 2022 22:07:41 GMT
- Title: Description-Driven Task-Oriented Dialog Modeling
- Authors: Jeffrey Zhao, Raghav Gupta, Yuan Cao, Dian Yu, Mingqiu Wang, Harrison
Lee, Abhinav Rastogi, Izhak Shafran, Yonghui Wu
- Abstract summary: We show that a language description-driven system exhibits better understanding of task specifications, higher performance on state tracking, improved data efficiency, and effective zero-shot transfer to unseen tasks.
We present a simple yet effective Description-Driven Dialog State Tracking (D3ST) model, which relies purely on schema descriptions and an "index-picking" mechanism.
- Score: 29.200221289845533
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Task-oriented dialogue (TOD) systems are required to identify key information
from conversations for the completion of given tasks. Such information is
conventionally specified in terms of intents and slots contained in
task-specific ontology or schemata. Since these schemata are designed by system
developers, the naming convention for slots and intents is not uniform across
tasks, and may not convey their semantics effectively. This can lead to models
memorizing arbitrary patterns in data, resulting in suboptimal performance and
generalization. In this paper, we propose that schemata should be modified by
replacing names or notations entirely with natural language descriptions. We
show that a language description-driven system exhibits better understanding of
task specifications, higher performance on state tracking, improved data
efficiency, and effective zero-shot transfer to unseen tasks. Following this
paradigm, we present a simple yet effective Description-Driven Dialog State
Tracking (D3ST) model, which relies purely on schema descriptions and an
"index-picking" mechanism. We demonstrate the superiority in quality, data
efficiency and robustness of our approach as measured on the MultiWOZ
(Budzianowski et al.,2018), SGD (Rastogi et al., 2020), and the recent SGD-X
(Lee et al., 2021) benchmarks.
Related papers
- Transforming Slot Schema Induction with Generative Dialogue State Inference [14.06505399101404]
Slot Induction (SSI) aims to automatically induce slots from unlabeled dialogue data.
Our SSI method discovers high-quality candidate information for representing dialogue state.
Experimental comparisons on the MultiWOZ and SGD datasets demonstrate that Generative Dialogue State Inference (GenDSI) outperforms the previous state-of-the-art.
arXiv Detail & Related papers (2024-08-03T02:41:10Z) - InstructTODS: Large Language Models for End-to-End Task-Oriented
Dialogue Systems [60.53276524369498]
Large language models (LLMs) have been used for diverse tasks in natural language processing (NLP)
We present InstructTODS, a novel framework for zero-shot end-to-end task-oriented dialogue systems.
InstructTODS generates a proxy belief state that seamlessly translates user intentions into dynamic queries.
arXiv Detail & Related papers (2023-10-13T06:36:26Z) - Slot Induction via Pre-trained Language Model Probing and Multi-level
Contrastive Learning [62.839109775887025]
Slot Induction (SI) task whose objective is to induce slot boundaries without explicit knowledge of token-level slot annotations.
We propose leveraging Unsupervised Pre-trained Language Model (PLM) Probing and Contrastive Learning mechanism to exploit unsupervised semantic knowledge extracted from PLM.
Our approach is shown to be effective in SI task and capable of bridging the gaps with token-level supervised models on two NLU benchmark datasets.
arXiv Detail & Related papers (2023-08-09T05:08:57Z) - Automated Few-shot Classification with Instruction-Finetuned Language
Models [76.69064714392165]
We show that AuT-Few outperforms state-of-the-art few-shot learning methods.
We also show that AuT-Few is the best ranking method across datasets on the RAFT few-shot benchmark.
arXiv Detail & Related papers (2023-05-21T21:50:27Z) - Zero-Shot Generalizable End-to-End Task-Oriented Dialog System using
Context Summarization and Domain Schema [2.7178968279054936]
State-of-the-art approaches in task-oriented dialog systems formulate the problem as a conditional sequence generation task.
This requires labeled training data for each new domain or task.
We introduce a novel Zero-Shot generalizable end-to-end Task-oriented Dialog system, ZS-ToD.
arXiv Detail & Related papers (2023-03-28T18:56:31Z) - AnyTOD: A Programmable Task-Oriented Dialog System [26.387269876540348]
We propose AnyTOD, an end-to-end task-oriented dialog (TOD) system with zero-shot capability for unseen tasks.
We demonstrate state-of-the-art results on the STAR and ABCD benchmarks, as well as AnyTOD's strong zero-shot transfer capability in low-resource settings.
arXiv Detail & Related papers (2022-12-20T01:23:01Z) - Discover, Explanation, Improvement: An Automatic Slice Detection
Framework for Natural Language Processing [72.14557106085284]
slice detection models (SDM) automatically identify underperforming groups of datapoints.
This paper proposes a benchmark named "Discover, Explain, improve (DEIM)" for classification NLP tasks.
Our evaluation shows that Edisa can accurately select error-prone datapoints with informative semantic features.
arXiv Detail & Related papers (2022-11-08T19:00:00Z) - A Multi-Task BERT Model for Schema-Guided Dialogue State Tracking [78.2700757742992]
Task-oriented dialogue systems often employ a Dialogue State Tracker (DST) to successfully complete conversations.
Recent state-of-the-art DST implementations rely on schemata of diverse services to improve model robustness.
We propose a single multi-task BERT-based model that jointly solves the three DST tasks of intent prediction, requested slot prediction and slot filling.
arXiv Detail & Related papers (2022-07-02T13:27:59Z) - Dialogue State Tracking with a Language Model using Schema-Driven
Prompting [18.83983018421701]
We introduce a new variation of the language modeling approach that uses schema-driven prompting to provide task-aware history encoding.
Our purely generative system achieves state-of-the-art performance on MultiWOZ 2.2 and achieves competitive performance on two other benchmarks: MultiWOZ 2.1 and M2M.
arXiv Detail & Related papers (2021-09-15T18:11:25Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z) - A Fast and Robust BERT-based Dialogue State Tracker for Schema-Guided
Dialogue Dataset [8.990035371365408]
We introduce FastSGT, a fast and robust BERT-based model for state tracking in goal-oriented dialogue systems.
The proposed model is designed for theGuided Dialogue dataset which contains natural language descriptions.
Our model keeps the efficiency in terms of computational and memory consumption while improving the accuracy significantly.
arXiv Detail & Related papers (2020-08-27T18:51:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.