Zero and Few-Shot Localization of Task-Oriented Dialogue Agents with a
Distilled Representation
- URL: http://arxiv.org/abs/2302.09424v1
- Date: Sat, 18 Feb 2023 21:30:36 GMT
- Title: Zero and Few-Shot Localization of Task-Oriented Dialogue Agents with a
Distilled Representation
- Authors: Mehrad Moradshahi, Sina J. Semnani, Monica S. Lam
- Abstract summary: We propose automatic methods that use ToD training data in a source language to build a high-quality functioning dialogue agent.
We show that our approach closes the accuracy gap between few-shot and existing full-shot methods for ToD agents.
- Score: 5.551814548069404
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Task-oriented Dialogue (ToD) agents are mostly limited to a few widely-spoken
languages, mainly due to the high cost of acquiring training data for each
language. Existing low-cost approaches that rely on cross-lingual embeddings or
naive machine translation sacrifice a lot of accuracy for data efficiency, and
largely fail in creating a usable dialogue agent. We propose automatic methods
that use ToD training data in a source language to build a high-quality
functioning dialogue agent in another target language that has no training data
(i.e. zero-shot) or a small training set (i.e. few-shot). Unlike most prior
work in cross-lingual ToD that only focuses on Dialogue State Tracking (DST),
we build an end-to-end agent.
We show that our approach closes the accuracy gap between few-shot and
existing full-shot methods for ToD agents. We achieve this by (1) improving the
dialogue data representation, (2) improving entity-aware machine translation,
and (3) automatic filtering of noisy translations.
We evaluate our approach on the recent bilingual dialogue dataset BiToD. In
Chinese to English transfer, in the zero-shot setting, our method achieves
46.7% and 22.0% in Task Success Rate (TSR) and Dialogue Success Rate (DSR)
respectively. In the few-shot setting where 10% of the data in the target
language is used, we improve the state-of-the-art by 15.2% and 14.0%, coming
within 5% of full-shot training.
Related papers
- Benchmarks Underestimate the Readiness of Multi-lingual Dialogue Agents [39.92509218078164]
We show for the first time, that in-context learning is sufficient to tackle multilingual TOD.
We test our approach on the multilingual TOD dataset X-RiSAWOZ, which has 12 domains in Chinese, English, French, Korean, Hindi, and code-mixed Hindi-English.
arXiv Detail & Related papers (2024-05-28T05:33:13Z) - Weakly Supervised Data Augmentation Through Prompting for Dialogue
Understanding [103.94325597273316]
We present a novel approach that iterates on augmentation quality by applying weakly-supervised filters.
We evaluate our methods on the emotion and act classification tasks in DailyDialog and the intent classification task in Facebook Multilingual Task-Oriented Dialogue.
For DailyDialog specifically, using 10% of the ground truth data we outperform the current state-of-the-art model which uses 100% of the data.
arXiv Detail & Related papers (2022-10-25T17:01:30Z) - Multi2WOZ: A Robust Multilingual Dataset and Conversational Pretraining
for Task-Oriented Dialog [67.20796950016735]
Multi2WOZ dataset spans four typologically diverse languages: Chinese, German, Arabic, and Russian.
We introduce a new framework for multilingual conversational specialization of pretrained language models (PrLMs) that aims to facilitate cross-lingual transfer for arbitrary downstream TOD tasks.
Our experiments show that, in most setups, the best performance entails the combination of (I) conversational specialization in the target language and (ii) few-shot transfer for the concrete TOD task.
arXiv Detail & Related papers (2022-05-20T18:35:38Z) - CrossAligner & Co: Zero-Shot Transfer Methods for Task-Oriented
Cross-lingual Natural Language Understanding [18.14437842819122]
CrossAligner is the principal method of a variety of effective approaches for zero-shot cross-lingual transfer.
We present a quantitative analysis of individual methods as well as their weighted combinations, several of which exceed state-of-the-art (SOTA) scores.
A detailed qualitative error analysis of the best methods shows that our fine-tuned language models can zero-shot transfer the task knowledge better than anticipated.
arXiv Detail & Related papers (2022-03-18T14:18:12Z) - Cross-Lingual Dialogue Dataset Creation via Outline-Based Generation [70.81596088969378]
Cross-lingual Outline-based Dialogue dataset (termed COD) enables natural language understanding.
COD enables dialogue state tracking, and end-to-end dialogue modelling and evaluation in 4 diverse languages.
arXiv Detail & Related papers (2022-01-31T18:11:21Z) - Contextual Semantic Parsing for Multilingual Task-Oriented Dialogues [7.8378818005171125]
Given a large-scale dialogue data set in one language, we can automatically produce an effective semantic for other languages using machine translation.
We propose automatic translation of dialogue datasets with alignment to ensure faithful translation of slot values.
We show that the succinct representation reduces the compounding effect of translation errors.
arXiv Detail & Related papers (2021-11-04T01:08:14Z) - Cross-lingual Intermediate Fine-tuning improves Dialogue State Tracking [84.50302759362698]
We enhance the transfer learning process by intermediate fine-tuning of pretrained multilingual models.
We use parallel and conversational movie subtitles datasets to design cross-lingual intermediate tasks.
We achieve impressive improvements (> 20% on goal accuracy) on the parallel MultiWoZ dataset and Multilingual WoZ dataset.
arXiv Detail & Related papers (2021-09-28T11:22:38Z) - BiToD: A Bilingual Multi-Domain Dataset For Task-Oriented Dialogue
Modeling [52.99188200886738]
BiToD is the first bilingual multi-domain dataset for end-to-end task-oriented dialogue modeling.
BiToD contains over 7k multi-domain dialogues (144k utterances) with a large and realistic bilingual knowledge base.
arXiv Detail & Related papers (2021-06-05T03:38:42Z) - Multilingual Speech Translation with Efficient Finetuning of Pretrained
Models [82.22294901727933]
A minimalistic LNA (LayerNorm and Attention) finetuning can achieve zero-shot crosslingual and cross-modality transfer ability.
Our approach demonstrates strong zero-shot performance in a many-to-many multilingual model.
arXiv Detail & Related papers (2020-10-24T08:15:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.