Polyglot Prompt: Multilingual Multitask PrompTraining
- URL: http://arxiv.org/abs/2204.14264v1
- Date: Fri, 29 Apr 2022 17:40:50 GMT
- Title: Polyglot Prompt: Multilingual Multitask PrompTraining
- Authors: Jinlan Fu, See-Kiong Ng, Pengfei Liu
- Abstract summary: This paper aims for a potential architectural breakthrough for multilingual learning and asks: could different tasks from different languages be modeled in a monolithic framework (without any task/language-specific module)?
We approach this goal by developing a learning framework Polyglot Prompt, where prompting methods are introduced to learn a unified semantic space for different languages and tasks after proper multilingual prompt engineering.
- Score: 35.70124413465395
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper aims for a potential architectural breakthrough for multilingual
learning and asks: could different tasks from different languages be modeled in
a monolithic framework (without any task/language-specific module)? The benefit
of achieving this is not only that systems trained on low resources scenario
can be assisted by more other languages and tasks, but opening new doors for
future multilingual research. We approach this goal by developing a learning
framework Polyglot Prompt, where prompting methods are introduced to learn a
unified semantic space for different languages and tasks after proper
multilingual prompt engineering. Experimentally, we perform a comprehensive
evaluation on 6 tasks (topic classification, sentiment classification, named
entity recognition, question answering, natural language inference,
summarization), 24 datasets, and 49 languages, which shows the efficacy of
multilingual multitask prompting training and suggests several interesting
observations. e.g., English prompts are polyglots since directly applying them
to task samples in other languages could result in a better improvement. We
also present an interpretable multilingual evaluation methodology and show how
the proposed framework, multilingual multitask prompt training, works. We
release all datasets prompted in the best setting and will release our code
soon.
Related papers
- Linguistically-Informed Multilingual Instruction Tuning: Is There an Optimal Set of Languages to Tune? [0.0]
This study proposes a method to select languages for instruction tuning in a linguistically informed way.
We use a simple algorithm to choose diverse languages and test their effectiveness on various benchmarks and open-ended questions.
Our results show that this careful selection generally leads to better outcomes than choosing languages at random.
arXiv Detail & Related papers (2024-10-10T10:57:24Z) - How do Large Language Models Handle Multilingualism? [81.15060972112563]
This study explores how large language models (LLMs) handle multilingualism.
LLMs initially understand the query, converting multilingual inputs into English for task-solving.
In the intermediate layers, they employ English for thinking and incorporate multilingual knowledge with self-attention and feed-forward structures.
arXiv Detail & Related papers (2024-02-29T02:55:26Z) - On Efficiently Acquiring Annotations for Multilingual Models [12.304046317362792]
We show that the strategy of joint learning across multiple languages using a single model performs substantially better than the aforementioned alternatives.
We show that this simple approach enables the model to be data efficient by allowing it to arbitrate its annotation budget to query languages it is less certain on.
arXiv Detail & Related papers (2022-04-03T07:42:13Z) - Towards Developing a Multilingual and Code-Mixed Visual Question
Answering System by Knowledge Distillation [20.33235443471006]
We propose a knowledge distillation approach to extend an English language-vision model (teacher) into an equally effective multilingual and code-mixed model (student)
We also create the large-scale multilingual and code-mixed VQA dataset in eleven different language setups.
Experimental results and in-depth analysis show the effectiveness of the proposed VQA model over the pre-trained language-vision models on eleven diverse language setups.
arXiv Detail & Related papers (2021-09-10T03:47:29Z) - Are Multilingual Models Effective in Code-Switching? [57.78477547424949]
We study the effectiveness of multilingual language models to understand their capability and adaptability to the mixed-language setting.
Our findings suggest that pre-trained multilingual models do not necessarily guarantee high-quality representations on code-switching.
arXiv Detail & Related papers (2021-03-24T16:20:02Z) - Meta-Learning for Effective Multi-task and Multilingual Modelling [23.53779501937046]
We propose a meta-learning approach to learn the interactions between both tasks and languages.
We present experiments on five different tasks and six different languages from the XTREME multilingual benchmark dataset.
arXiv Detail & Related papers (2021-01-25T19:30:26Z) - CoSDA-ML: Multi-Lingual Code-Switching Data Augmentation for Zero-Shot
Cross-Lingual NLP [68.2650714613869]
We propose a data augmentation framework to generate multi-lingual code-switching data to fine-tune mBERT.
Compared with the existing work, our method does not rely on bilingual sentences for training, and requires only one training process for multiple target languages.
arXiv Detail & Related papers (2020-06-11T13:15:59Z) - M3P: Learning Universal Representations via Multitask Multilingual
Multimodal Pre-training [119.16007395162431]
M3P is a Multilingual Multimodal Pre-trained model that combines multilingual pre-training and multimodal pre-training.
We show that M3P can achieve comparable results for English and new state-of-the-art results for non-English languages.
arXiv Detail & Related papers (2020-06-04T03:54:29Z) - Learning to Scale Multilingual Representations for Vision-Language Tasks [51.27839182889422]
The effectiveness of SMALR is demonstrated with ten diverse languages, over twice the number supported in vision-language tasks to date.
We evaluate on multilingual image-sentence retrieval and outperform prior work by 3-4% with less than 1/5th the training parameters compared to other word embedding methods.
arXiv Detail & Related papers (2020-04-09T01:03:44Z) - XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating
Cross-lingual Generalization [128.37244072182506]
Cross-lingual TRansfer Evaluation of Multilinguals XTREME is a benchmark for evaluating the cross-lingual generalization capabilities of multilingual representations across 40 languages and 9 tasks.
We demonstrate that while models tested on English reach human performance on many tasks, there is still a sizable gap in the performance of cross-lingually transferred models.
arXiv Detail & Related papers (2020-03-24T19:09:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.