Zero-shot Generalization in Dialog State Tracking through Generative
Question Answering
- URL: http://arxiv.org/abs/2101.08333v1
- Date: Wed, 20 Jan 2021 21:47:20 GMT
- Title: Zero-shot Generalization in Dialog State Tracking through Generative
Question Answering
- Authors: Shuyang Li, Jin Cao, Mukund Sridhar, Henghui Zhu, Shang-Wen Li, Wael
Hamza, Julian McAuley
- Abstract summary: We introduce a novel framework that supports natural language queries for unseen constraints and slots in task-oriented dialogs.
Our approach is based on generative question-answering using a conditional domain model pre-trained on substantive English sentences.
- Score: 10.81203437307028
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dialog State Tracking (DST), an integral part of modern dialog systems, aims
to track user preferences and constraints (slots) in task-oriented dialogs. In
real-world settings with constantly changing services, DST systems must
generalize to new domains and unseen slot types. Existing methods for DST do
not generalize well to new slot names and many require known ontologies of slot
types and values for inference. We introduce a novel ontology-free framework
that supports natural language queries for unseen constraints and slots in
multi-domain task-oriented dialogs. Our approach is based on generative
question-answering using a conditional language model pre-trained on
substantive English sentences. Our model improves joint goal accuracy in
zero-shot domain adaptation settings by up to 9% (absolute) over the previous
state-of-the-art on the MultiWOZ 2.1 dataset.
Related papers
- A Zero-Shot Open-Vocabulary Pipeline for Dialogue Understanding [0.0]
We propose a zero-shot, open-vocabulary system that integrates domain classification and State Tracking (DST) in a single pipeline.
Our approach includes reformulating DST as a question-answering task for less capable models and employing self-refining prompts for more adaptable ones.
arXiv Detail & Related papers (2024-09-24T08:33:41Z) - DiSTRICT: Dialogue State Tracking with Retriever Driven In-Context
Tuning [7.5700317050237365]
We propose DiSTRICT, a generalizable in-context tuning approach for Dialogue State Tracking (DST)
DSTRICT retrieves highly relevant training examples for a given dialogue to fine-tune the model without any hand-crafted templates.
Experiments with the MultiWOZ benchmark datasets show that DiSTRICT outperforms existing approaches in various zero-shot and few-shot settings.
arXiv Detail & Related papers (2022-12-06T09:40:15Z) - A Slot Is Not Built in One Utterance: Spoken Language Dialogs with
Sub-Slots [67.69407159704328]
This paper defines a new task named Sub-Slot based Task-Oriented Dialog (SSTOD)
The dataset includes a total of 40K dialogs and 500K utterances from four different domains: Chinese names, phone numbers, ID numbers and license plate numbers.
We find some new linguistic phenomena and interactive manners in SSTOD which raise critical challenges of building dialog agents for the task.
arXiv Detail & Related papers (2022-03-21T07:10:19Z) - In-Context Learning for Few-Shot Dialogue State Tracking [55.91832381893181]
We propose an in-context (IC) learning framework for few-shot dialogue state tracking (DST)
A large pre-trained language model (LM) takes a test instance and a few annotated examples as input, and directly decodes the dialogue states without any parameter updates.
This makes the LM more flexible and scalable compared to prior few-shot DST work when adapting to new domains and scenarios.
arXiv Detail & Related papers (2022-03-16T11:58:24Z) - Cross-Lingual Dialogue Dataset Creation via Outline-Based Generation [70.81596088969378]
Cross-lingual Outline-based Dialogue dataset (termed COD) enables natural language understanding.
COD enables dialogue state tracking, and end-to-end dialogue modelling and evaluation in 4 diverse languages.
arXiv Detail & Related papers (2022-01-31T18:11:21Z) - Zero-Shot Dialogue State Tracking via Cross-Task Transfer [69.70718906395182]
We propose to transfer the textitcross-task knowledge from general question answering (QA) corpora for the zero-shot dialogue state tracking task.
Specifically, we propose TransferQA, a transferable generative QA model that seamlessly combines extractive QA and multi-choice QA.
In addition, we introduce two effective ways to construct unanswerable questions, namely, negative question sampling and context truncation.
arXiv Detail & Related papers (2021-09-10T03:57:56Z) - CREDIT: Coarse-to-Fine Sequence Generation for Dialogue State Tracking [44.38388988238695]
A dialogue state tracker aims to accurately find a compact representation of the current dialogue status.
We employ a structured state representation and cast dialogue state tracking as a sequence generation problem.
Experiments demonstrate our tracker achieves encouraging joint goal accuracy for the five domains in MultiWOZ 2.0 and MultiWOZ 2.1 datasets.
arXiv Detail & Related papers (2020-09-22T10:27:18Z) - Few-shot Natural Language Generation for Task-Oriented Dialog [113.07438787659859]
We present FewShotWoz, the first NLG benchmark to simulate the few-shot learning setting in task-oriented dialog systems.
We develop the SC-GPT model, which is pre-trained on a large set of annotated NLG corpus to acquire the controllable generation ability.
Experiments on FewShotWoz and the large Multi-Domain-WOZ datasets show that the proposed SC-GPT significantly outperforms existing methods.
arXiv Detail & Related papers (2020-02-27T18:48:33Z) - Non-Autoregressive Dialog State Tracking [122.2328875457225]
We propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST)
NADST can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots.
Our results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus.
arXiv Detail & Related papers (2020-02-19T06:39:26Z) - MA-DST: Multi-Attention Based Scalable Dialog State Tracking [13.358314140896937]
Dialog State Tracking dialog agents provide a natural language interface for users to complete their goal.
To enable accurate multi-domain DST, the model needs to encode dependencies between past utterances and slot semantics.
We introduce a novel architecture for this task to encode the conversation history and slot semantics.
arXiv Detail & Related papers (2020-02-07T05:34:58Z) - Domain-Aware Dialogue State Tracker for Multi-Domain Dialogue Systems [2.3859169601259347]
In task-oriented dialogue systems the dialogue state tracker (DST) component is responsible for predicting the state of the dialogue based on the dialogue history.
We propose a domain-aware dialogue state tracker that is completely data-driven and it is modeled to predict for dynamic service schemas.
arXiv Detail & Related papers (2020-01-21T13:41:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.