XPersona: Evaluating Multilingual Personalized Chatbot
- URL: http://arxiv.org/abs/2003.07568v2
- Date: Wed, 8 Apr 2020 07:38:28 GMT
- Title: XPersona: Evaluating Multilingual Personalized Chatbot
- Authors: Zhaojiang Lin, Zihan Liu, Genta Indra Winata, Samuel Cahyawijaya,
Andrea Madotto, Yejin Bang, Etsuko Ishii, Pascale Fung
- Abstract summary: We propose a multi-lingual extension of Persona-Chat, namely XPersona.
Our dataset includes persona conversations in six different languages other than English for building and evaluating multilingual personalized agents.
- Score: 76.00426517401894
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Personalized dialogue systems are an essential step toward better
human-machine interaction. Existing personalized dialogue agents rely on
properly designed conversational datasets, which are mostly monolingual (e.g.,
English), which greatly limits the usage of conversational agents in other
languages. In this paper, we propose a multi-lingual extension of Persona-Chat,
namely XPersona. Our dataset includes persona conversations in six different
languages other than English for building and evaluating multilingual
personalized agents. We experiment with both multilingual and cross-lingual
trained baselines, and evaluate them against monolingual and
translation-pipeline models using both automatic and human evaluation.
Experimental results show that the multilingual trained models outperform the
translation-pipeline and that they are on par with the monolingual models, with
the advantage of having a single model across multiple languages. On the other
hand, the state-of-the-art cross-lingual trained models achieve inferior
performance to the other models, showing that cross-lingual conversation
modeling is a challenging task. We hope that our dataset and baselines will
accelerate research in multilingual dialogue systems.
Related papers
- Multilingual Turn-taking Prediction Using Voice Activity Projection [25.094622033971643]
This paper investigates the application of voice activity projection (VAP), a predictive turn-taking model for spoken dialogue, on multilingual data.
The results show that a monolingual VAP model trained on one language does not make good predictions when applied to other languages.
A multilingual model, trained on all three languages, demonstrates predictive performance on par with monolingual models across all languages.
arXiv Detail & Related papers (2024-03-11T07:50:29Z) - Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of
Multilingual Language Models [73.11488464916668]
This study investigates the dynamics of the multilingual pretraining process.
We probe checkpoints taken from throughout XLM-R pretraining, using a suite of linguistic tasks.
Our analysis shows that the model achieves high in-language performance early on, with lower-level linguistic skills acquired before more complex ones.
arXiv Detail & Related papers (2022-05-24T03:35:00Z) - Exploring Teacher-Student Learning Approach for Multi-lingual
Speech-to-Intent Classification [73.5497360800395]
We develop an end-to-end system that supports multiple languages.
We exploit knowledge from a pre-trained multi-lingual natural language processing model.
arXiv Detail & Related papers (2021-09-28T04:43:11Z) - Discovering Representation Sprachbund For Multilingual Pre-Training [139.05668687865688]
We generate language representation from multilingual pre-trained models and conduct linguistic analysis.
We cluster all the target languages into multiple groups and name each group as a representation sprachbund.
Experiments are conducted on cross-lingual benchmarks and significant improvements are achieved compared to strong baselines.
arXiv Detail & Related papers (2021-09-01T09:32:06Z) - How Good is Your Tokenizer? On the Monolingual Performance of
Multilingual Language Models [96.32118305166412]
We study a set of nine typologically diverse languages with readily available pretrained monolingual models on a set of five diverse monolingual downstream tasks.
We find that languages which are adequately represented in the multilingual model's vocabulary exhibit negligible performance decreases over their monolingual counterparts.
arXiv Detail & Related papers (2020-12-31T14:11:00Z) - Evaluating Cross-Lingual Transfer Learning Approaches in Multilingual
Conversational Agent Models [1.52292571922932]
We propose a general multilingual model framework for Natural Language Understanding (NLU) models.
We show that these multilingual models can reach same or better performance compared to monolingual models across language-specific test data.
arXiv Detail & Related papers (2020-12-07T17:14:52Z) - Towards Fully Bilingual Deep Language Modeling [1.3455090151301572]
We consider whether it is possible to pre-train a bilingual model for two remotely related languages without compromising performance at either language.
We create a Finnish-English bilingual BERT model and evaluate its performance on datasets used to evaluate the corresponding monolingual models.
Our bilingual model performs on par with Google's original English BERT on GLUE and nearly matches the performance of monolingual Finnish BERT on a range of Finnish NLP tasks.
arXiv Detail & Related papers (2020-10-22T12:22:50Z) - Multilingual Translation with Extensible Multilingual Pretraining and
Finetuning [77.33262578776291]
Previous work has demonstrated that machine translation systems can be created by finetuning on bitext.
We show that multilingual translation models can be created through multilingual finetuning.
We demonstrate that pretrained models can be extended to incorporate additional languages without loss of performance.
arXiv Detail & Related papers (2020-08-02T05:36:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.