Schema Encoding for Transferable Dialogue State Tracking
- URL: http://arxiv.org/abs/2210.02351v1
- Date: Wed, 5 Oct 2022 15:53:06 GMT
- Title: Schema Encoding for Transferable Dialogue State Tracking
- Authors: Hyunmin Jeon and Gary Geunbae Lee
- Abstract summary: Dialogue state tracking (DST) is an essential sub-task for task-oriented dialogue systems.
Recent work has focused on deep neural models for DST.
Applying them to another domain needs a new dataset.
- Score: 2.5838973036257458
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Dialogue state tracking (DST) is an essential sub-task for task-oriented
dialogue systems. Recent work has focused on deep neural models for DST.
However, the neural models require a large dataset for training. Furthermore,
applying them to another domain needs a new dataset because the neural models
are generally trained to imitate the given dataset. In this paper, we propose
Schema Encoding for Transferable Dialogue State Tracking (SETDST), which is a
neural DST method for effective transfer to new domains. Transferable DST could
assist developments of dialogue systems even with few dataset on target
domains. We use a schema encoder not just to imitate the dataset but to
comprehend the schema of the dataset. We aim to transfer the model to new
domains by encoding new schemas and using them for DST on multi-domain
settings. As a result, SET-DST improved the joint accuracy by 1.46 points on
MultiWOZ 2.1.
Related papers
- UNO-DST: Leveraging Unlabelled Data in Zero-Shot Dialogue State Tracking [54.51316566989655]
Previous zero-shot dialogue state tracking (DST) methods only apply transfer learning, ignoring unlabelled data in the target domain.
We transform zero-shot DST into few-shot DST by utilising such unlabelled data via joint and self-training methods.
We demonstrate this method's effectiveness on general language models in zero-shot scenarios, improving average joint goal accuracy by 8% across all domains in MultiWOZ.
arXiv Detail & Related papers (2023-10-16T15:16:16Z) - Choice Fusion as Knowledge for Zero-Shot Dialogue State Tracking [5.691339955497443]
zero-shot dialogue state tracking (DST) tracks user's requirements in task-oriented dialogues without training on desired domains.
We propose CoFunDST, which is trained on domain-agnostic QA datasets and directly uses candidate choices of slot-values as knowledge for zero-shot dialogue-state generation.
Our proposed model achieves outperformed joint goal accuracy compared to existing zero-shot DST approaches in most domains on the MultiWOZ 2.1.
arXiv Detail & Related papers (2023-02-25T07:32:04Z) - StorSeismic: A new paradigm in deep learning for seismic processing [0.0]
StorSeismic is a framework for seismic data processing.
We pre-train seismic data, along with synthetically generated ones, in the self-supervised step.
Then, we use the labeled synthetic data to fine-tune the pre-trained network in a supervised fashion to perform various seismic processing tasks.
arXiv Detail & Related papers (2022-04-30T09:55:00Z) - Prompt Learning for Few-Shot Dialogue State Tracking [75.50701890035154]
This paper focuses on how to learn a dialogue state tracking (DST) model efficiently with limited labeled data.
We design a prompt learning framework for few-shot DST, which consists of two main components: value-based prompt and inverse prompt mechanism.
Experiments show that our model can generate unseen slots and outperforms existing state-of-the-art few-shot methods.
arXiv Detail & Related papers (2022-01-15T07:37:33Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Domain Adaptation in Dialogue Systems using Transfer and Meta-Learning [12.64591916699374]
Current generative-based dialogue systems fail to adapt to new unseen domains when only a small amount of target data is available.
We propose a method that adapts to unseen domains by combining both transfer and meta-learning.
arXiv Detail & Related papers (2021-02-22T16:16:57Z) - Few Shot Dialogue State Tracking using Meta-learning [3.6292310166028403]
Dialogue State Tracking (DST) forms a core component of automated systems designed for specific goals like hotel, taxi reservation, tourist information, etc.
With the increasing need to deploy such systems in new domains, solving the problem of zero/few-shot DST has become necessary.
Our proposed meta-learner is agnostic of the underlying model and hence any existing state-of-the-art DST system can improve its performance on unknown domains using our training strategy.
arXiv Detail & Related papers (2021-01-17T20:47:06Z) - Improving Limited Labeled Dialogue State Tracking with Self-Supervision [91.68515201803986]
Existing dialogue state tracking (DST) models require plenty of labeled data.
We present and investigate two self-supervised objectives: preserving latent consistency and modeling conversational behavior.
Our proposed self-supervised signals can improve joint goal accuracy by 8.95% when only 1% labeled data is used.
arXiv Detail & Related papers (2020-10-26T21:57:42Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Few-shot Natural Language Generation for Task-Oriented Dialog [113.07438787659859]
We present FewShotWoz, the first NLG benchmark to simulate the few-shot learning setting in task-oriented dialog systems.
We develop the SC-GPT model, which is pre-trained on a large set of annotated NLG corpus to acquire the controllable generation ability.
Experiments on FewShotWoz and the large Multi-Domain-WOZ datasets show that the proposed SC-GPT significantly outperforms existing methods.
arXiv Detail & Related papers (2020-02-27T18:48:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.