A Fast and Robust BERT-based Dialogue State Tracker for Schema-Guided
Dialogue Dataset
- URL: http://arxiv.org/abs/2008.12335v1
- Date: Thu, 27 Aug 2020 18:51:18 GMT
- Title: A Fast and Robust BERT-based Dialogue State Tracker for Schema-Guided
Dialogue Dataset
- Authors: Vahid Noroozi, Yang Zhang, Evelina Bakhturina, Tomasz Kornuta
- Abstract summary: We introduce FastSGT, a fast and robust BERT-based model for state tracking in goal-oriented dialogue systems.
The proposed model is designed for theGuided Dialogue dataset which contains natural language descriptions.
Our model keeps the efficiency in terms of computational and memory consumption while improving the accuracy significantly.
- Score: 8.990035371365408
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dialog State Tracking (DST) is one of the most crucial modules for
goal-oriented dialogue systems. In this paper, we introduce FastSGT (Fast
Schema Guided Tracker), a fast and robust BERT-based model for state tracking
in goal-oriented dialogue systems. The proposed model is designed for the
Schema-Guided Dialogue (SGD) dataset which contains natural language
descriptions for all the entities including user intents, services, and slots.
The model incorporates two carry-over procedures for handling the extraction of
the values not explicitly mentioned in the current user utterance. It also uses
multi-head attention projections in some of the decoders to have a better
modelling of the encoder outputs. In the conducted experiments we compared
FastSGT to the baseline model for the SGD dataset. Our model keeps the
efficiency in terms of computational and memory consumption while improving the
accuracy significantly. Additionally, we present ablation studies measuring the
impact of different parts of the model on its performance. We also show the
effectiveness of data augmentation for improving the accuracy without
increasing the amount of computational resources.
Related papers
- Plan, Generate and Complicate: Improving Low-resource Dialogue State Tracking via Easy-to-Difficult Zero-shot Data Augmentation [5.042738414157664]
We propose EDZ-DA, an Easy-to-Difficult Zero-shot Data Augmentation framework for low-resource dialogue state tracking.
We also complicate the dialogues based on the domain relation to enhance the model's capability for co-reference slot tracking.
arXiv Detail & Related papers (2024-06-13T06:49:03Z) - TOD-Flow: Modeling the Structure of Task-Oriented Dialogues [77.15457469745364]
We propose a novel approach focusing on inferring the TOD-Flow graph from dialogue data annotated with dialog acts.
The inferred TOD-Flow graph can be easily integrated with any dialogue model to improve its prediction performance, transparency, and controllability.
arXiv Detail & Related papers (2023-12-07T20:06:23Z) - Injecting linguistic knowledge into BERT for Dialogue State Tracking [60.42231674887294]
This paper proposes a method that extracts linguistic knowledge via an unsupervised framework.
We then utilize this knowledge to augment BERT's performance and interpretability in Dialogue State Tracking (DST) tasks.
We benchmark this framework on various DST tasks and observe a notable improvement in accuracy.
arXiv Detail & Related papers (2023-11-27T08:38:42Z) - Unlocking the Potential of User Feedback: Leveraging Large Language
Model as User Simulator to Enhance Dialogue System [65.93577256431125]
We propose an alternative approach called User-Guided Response Optimization (UGRO) to combine it with a smaller task-oriented dialogue model.
This approach uses LLM as annotation-free user simulator to assess dialogue responses, combining them with smaller fine-tuned end-to-end TOD models.
Our approach outperforms previous state-of-the-art (SOTA) results.
arXiv Detail & Related papers (2023-06-16T13:04:56Z) - Span-Selective Linear Attention Transformers for Effective and Robust
Schema-Guided Dialogue State Tracking [7.176787451868171]
We introduce SPLAT, a novel architecture which achieves better generalization and efficiency than prior approaches.
We demonstrate the effectiveness of our model on theGuided Dialogue (SGD) and MultiWOZ datasets.
arXiv Detail & Related papers (2023-06-15T17:59:31Z) - TLAG: An Informative Trigger and Label-Aware Knowledge Guided Model for
Dialogue-based Relation Extraction [43.98737106776455]
Dialogue-based Relation Extraction (DRE) aims to predict the relation type of argument pairs that are mentioned in dialogue.
We propose TLAG, which fully leverages the trigger and label-aware knowledge to guide the relation extraction.
Experimental results on the DialogRE dataset show that our TLAG outperforms the baseline models.
arXiv Detail & Related papers (2023-03-30T03:10:28Z) - A Multi-Task BERT Model for Schema-Guided Dialogue State Tracking [78.2700757742992]
Task-oriented dialogue systems often employ a Dialogue State Tracker (DST) to successfully complete conversations.
Recent state-of-the-art DST implementations rely on schemata of diverse services to improve model robustness.
We propose a single multi-task BERT-based model that jointly solves the three DST tasks of intent prediction, requested slot prediction and slot filling.
arXiv Detail & Related papers (2022-07-02T13:27:59Z) - Prompt Learning for Few-Shot Dialogue State Tracking [75.50701890035154]
This paper focuses on how to learn a dialogue state tracking (DST) model efficiently with limited labeled data.
We design a prompt learning framework for few-shot DST, which consists of two main components: value-based prompt and inverse prompt mechanism.
Experiments show that our model can generate unseen slots and outperforms existing state-of-the-art few-shot methods.
arXiv Detail & Related papers (2022-01-15T07:37:33Z) - SGD-QA: Fast Schema-Guided Dialogue State Tracking for Unseen Services [15.21976869687864]
We propose SGD-QA, a model for schema-guided dialogue state tracking based on a question answering approach.
The proposed multi-pass model shares a single encoder between the domain information and dialogue utterance.
The model improves performance on unseen services by at least 1.6x compared to single-pass baseline models.
arXiv Detail & Related papers (2021-05-17T17:54:32Z) - Improving Limited Labeled Dialogue State Tracking with Self-Supervision [91.68515201803986]
Existing dialogue state tracking (DST) models require plenty of labeled data.
We present and investigate two self-supervised objectives: preserving latent consistency and modeling conversational behavior.
Our proposed self-supervised signals can improve joint goal accuracy by 8.95% when only 1% labeled data is used.
arXiv Detail & Related papers (2020-10-26T21:57:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.