Zero-Shot Transfer Learning with Synthesized Data for Multi-Domain
Dialogue State Tracking
- URL: http://arxiv.org/abs/2005.00891v1
- Date: Sat, 2 May 2020 18:00:48 GMT
- Title: Zero-Shot Transfer Learning with Synthesized Data for Multi-Domain
Dialogue State Tracking
- Authors: Giovanni Campagna and Agata Foryciarz and Mehrad Moradshahi and Monica
S. Lam
- Abstract summary: We show that data augmentation through synthesized data can improve the accuracy of zero-shot learning.
We improve the zero-shot learning state of the art on average across domains by 21%.
- Score: 8.151397072537797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Zero-shot transfer learning for multi-domain dialogue state tracking can
allow us to handle new domains without incurring the high cost of data
acquisition. This paper proposes new zero-short transfer learning technique for
dialogue state tracking where the in-domain training data are all synthesized
from an abstract dialogue model and the ontology of the domain. We show that
data augmentation through synthesized data can improve the accuracy of
zero-shot learning for both the TRADE model and the BERT-based SUMBT model on
the MultiWOZ 2.1 dataset. We show training with only synthesized in-domain data
on the SUMBT model can reach about 2/3 of the accuracy obtained with the full
training dataset. We improve the zero-shot learning state of the art on average
across domains by 21%.
Related papers
- Diverse and Effective Synthetic Data Generation for Adaptable Zero-Shot Dialogue State Tracking [12.116834890063146]
We show substantial performance gains in zero-shot dialogue state tracking (DST) by enhancing training data diversity through synthetic data generation.
Existing DST datasets are severely limited in the number of application domains and slot types they cover due to the high costs of data collection.
This work addresses this challenge with a novel, fully automatic data generation approach that creates synthetic zero-shot DST datasets.
arXiv Detail & Related papers (2024-05-21T03:04:14Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Training on Synthetic Data Beats Real Data in Multimodal Relation
Extraction [8.038421100401132]
In this paper, we consider a novel problem setting, where only unimodal data, either text or image, are available during training.
We aim to train a multimodal relation from synthetic data that perform well on real multimodal test data.
Our best model trained on completely synthetic images outperforms prior state-of-the-art models trained on real multimodal data by a margin of 3.76% in F1.
arXiv Detail & Related papers (2023-12-05T08:11:34Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Domain Adaptation in Dialogue Systems using Transfer and Meta-Learning [12.64591916699374]
Current generative-based dialogue systems fail to adapt to new unseen domains when only a small amount of target data is available.
We propose a method that adapts to unseen domains by combining both transfer and meta-learning.
arXiv Detail & Related papers (2021-02-22T16:16:57Z) - Dual-Teacher++: Exploiting Intra-domain and Inter-domain Knowledge with
Reliable Transfer for Cardiac Segmentation [69.09432302497116]
We propose a cutting-edge semi-supervised domain adaptation framework, namely Dual-Teacher++.
We design novel dual teacher models, including an inter-domain teacher model to explore cross-modality priors from source domain (e.g., MR) and an intra-domain teacher model to investigate the knowledge beneath unlabeled target domain.
In this way, the student model can obtain reliable dual-domain knowledge and yield improved performance on target domain data.
arXiv Detail & Related papers (2021-01-07T05:17:38Z) - Improving Zero and Few-Shot Abstractive Summarization with Intermediate
Fine-tuning and Data Augmentation [101.26235068460551]
Models pretrained with self-supervised objectives on large text corpora achieve state-of-the-art performance on English text summarization tasks.
Models are typically fine-tuned on hundreds of thousands of data points, an infeasible requirement when applying summarization to new, niche domains.
We introduce a novel and generalizable method, called WikiTransfer, for fine-tuning pretrained models for summarization in an unsupervised, dataset-specific manner.
arXiv Detail & Related papers (2020-10-24T08:36:49Z) - Multi-Domain Spoken Language Understanding Using Domain- and Task-Aware
Parameterization [78.93669377251396]
Spoken language understanding has been addressed as a supervised learning problem, where a set of training data is available for each domain.
One existing approach solves the problem by conducting multi-domain learning, using shared parameters for joint training across domains.
We propose to improve the parameterization of this method by using domain-specific and task-specific model parameters.
arXiv Detail & Related papers (2020-04-30T15:15:40Z) - Dynamic Fusion Network for Multi-Domain End-to-end Task-Oriented Dialog [70.79442700890843]
We propose a novel Dynamic Fusion Network (DF-Net) which automatically exploit the relevance between the target domain and each domain.
With little training data, we show its transferability by outperforming prior best model by 13.9% on average.
arXiv Detail & Related papers (2020-04-23T08:17:22Z) - Abstractive Summarization for Low Resource Data using Domain Transfer
and Data Synthesis [1.148539813252112]
We explore using domain transfer and data synthesis to improve the performance of recent abstractive summarization methods.
We show that tuning state of the art model trained on newspaper data could boost performance on student reflection data.
We propose a template-based model to synthesize new data, which when incorporated into training further increased ROUGE scores.
arXiv Detail & Related papers (2020-02-09T17:49:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.