Meta-Learning for Effective Multi-task and Multilingual Modelling
- URL: http://arxiv.org/abs/2101.10368v2
- Date: Wed, 27 Jan 2021 16:35:02 GMT
- Title: Meta-Learning for Effective Multi-task and Multilingual Modelling
- Authors: Ishan Tarunesh, Sushil Khyalia, Vishwajeet Kumar, Ganesh Ramakrishnan,
Preethi Jyothi
- Abstract summary: We propose a meta-learning approach to learn the interactions between both tasks and languages.
We present experiments on five different tasks and six different languages from the XTREME multilingual benchmark dataset.
- Score: 23.53779501937046
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Natural language processing (NLP) tasks (e.g. question-answering in English)
benefit from knowledge of other tasks (e.g. named entity recognition in
English) and knowledge of other languages (e.g. question-answering in Spanish).
Such shared representations are typically learned in isolation, either across
tasks or across languages. In this work, we propose a meta-learning approach to
learn the interactions between both tasks and languages. We also investigate
the role of different sampling strategies used during meta-learning. We present
experiments on five different tasks and six different languages from the XTREME
multilingual benchmark dataset. Our meta-learned model clearly improves in
performance compared to competitive baseline models that also include
multi-task baselines. We also present zero-shot evaluations on unseen target
languages to demonstrate the utility of our proposed model.
Related papers
- SkillNet-X: A Multilingual Multitask Model with Sparsely Activated
Skills [51.74947795895178]
This paper proposes a general multilingual multitask model, named SkillNet-X.
We define several language-specific skills and task-specific skills, each of which corresponds to a skill module.
We evaluate SkillNet-X on eleven natural language understanding datasets in four languages.
arXiv Detail & Related papers (2023-06-28T12:53:30Z) - Persian Natural Language Inference: A Meta-learning approach [6.832341432995628]
This paper proposes a meta-learning approach for inferring natural language in Persian.
We evaluate the proposed method using four languages and an auxiliary task.
arXiv Detail & Related papers (2022-05-18T06:51:58Z) - Polyglot Prompt: Multilingual Multitask PrompTraining [35.70124413465395]
This paper aims for a potential architectural breakthrough for multilingual learning and asks: could different tasks from different languages be modeled in a monolithic framework (without any task/language-specific module)?
We approach this goal by developing a learning framework Polyglot Prompt, where prompting methods are introduced to learn a unified semantic space for different languages and tasks after proper multilingual prompt engineering.
arXiv Detail & Related papers (2022-04-29T17:40:50Z) - On Efficiently Acquiring Annotations for Multilingual Models [12.304046317362792]
We show that the strategy of joint learning across multiple languages using a single model performs substantially better than the aforementioned alternatives.
We show that this simple approach enables the model to be data efficient by allowing it to arbitrate its annotation budget to query languages it is less certain on.
arXiv Detail & Related papers (2022-04-03T07:42:13Z) - Are Multilingual Models Effective in Code-Switching? [57.78477547424949]
We study the effectiveness of multilingual language models to understand their capability and adaptability to the mixed-language setting.
Our findings suggest that pre-trained multilingual models do not necessarily guarantee high-quality representations on code-switching.
arXiv Detail & Related papers (2021-03-24T16:20:02Z) - CoSDA-ML: Multi-Lingual Code-Switching Data Augmentation for Zero-Shot
Cross-Lingual NLP [68.2650714613869]
We propose a data augmentation framework to generate multi-lingual code-switching data to fine-tune mBERT.
Compared with the existing work, our method does not rely on bilingual sentences for training, and requires only one training process for multiple target languages.
arXiv Detail & Related papers (2020-06-11T13:15:59Z) - Learning to Scale Multilingual Representations for Vision-Language Tasks [51.27839182889422]
The effectiveness of SMALR is demonstrated with ten diverse languages, over twice the number supported in vision-language tasks to date.
We evaluate on multilingual image-sentence retrieval and outperform prior work by 3-4% with less than 1/5th the training parameters compared to other word embedding methods.
arXiv Detail & Related papers (2020-04-09T01:03:44Z) - XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating
Cross-lingual Generalization [128.37244072182506]
Cross-lingual TRansfer Evaluation of Multilinguals XTREME is a benchmark for evaluating the cross-lingual generalization capabilities of multilingual representations across 40 languages and 9 tasks.
We demonstrate that while models tested on English reach human performance on many tasks, there is still a sizable gap in the performance of cross-lingually transferred models.
arXiv Detail & Related papers (2020-03-24T19:09:37Z) - Zero-Shot Cross-Lingual Transfer with Meta Learning [45.29398184889296]
We consider the setting of training models on multiple languages at the same time, when little or no data is available for languages other than English.
We show that this challenging setup can be approached using meta-learning.
We experiment using standard supervised, zero-shot cross-lingual, as well as few-shot cross-lingual settings for different natural language understanding tasks.
arXiv Detail & Related papers (2020-03-05T16:07:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.