Discovering Language-neutral Sub-networks in Multilingual Language
Models
- URL: http://arxiv.org/abs/2205.12672v1
- Date: Wed, 25 May 2022 11:35:41 GMT
- Title: Discovering Language-neutral Sub-networks in Multilingual Language
Models
- Authors: Negar Foroutan, Mohammadreza Banaei, Remi Lebret, Antoine Bosselut,
Karl Aberer
- Abstract summary: Language neutrality of multilingual models is a function of the overlap between language-encoding sub-networks of these models.
Using mBERT as a foundation, we employ the lottery ticket hypothesis to discover sub-networks that are individually optimized for various languages and tasks.
We conclude that mBERT is comprised of a language-neutral sub-network shared among many languages, along with multiple ancillary language-specific sub-networks.
- Score: 15.94622051535847
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multilingual pre-trained language models perform remarkably well on
cross-lingual transfer for downstream tasks. Despite their impressive
performance, our understanding of their language neutrality (i.e., the extent
to which they use shared representations to encode similar phenomena across
languages) and its role in achieving such performance remain open questions. In
this work, we conceptualize language neutrality of multilingual models as a
function of the overlap between language-encoding sub-networks of these models.
Using mBERT as a foundation, we employ the lottery ticket hypothesis to
discover sub-networks that are individually optimized for various languages and
tasks. Using three distinct tasks and eleven typologically-diverse languages in
our evaluation, we show that the sub-networks found for different languages are
in fact quite similar, supporting the idea that mBERT jointly encodes multiple
languages in shared parameters. We conclude that mBERT is comprised of a
language-neutral sub-network shared among many languages, along with multiple
ancillary language-specific sub-networks, with the former playing a more
prominent role in mBERT's impressive cross-lingual performance.
Related papers
- xCoT: Cross-lingual Instruction Tuning for Cross-lingual
Chain-of-Thought Reasoning [36.34986831526529]
Chain-of-thought (CoT) has emerged as a powerful technique to elicit reasoning in large language models.
We propose a cross-lingual instruction fine-tuning framework (xCOT) to transfer knowledge from high-resource languages to low-resource languages.
arXiv Detail & Related papers (2024-01-13T10:53:53Z) - Establishing Interlingua in Multilingual Language Models [0.0]
We show that different languages do converge to a shared space in large multilingual language models.
We extend our analysis to 28 diverse languages and find that the interlingual space exhibits a particular structure similar to the linguistic relatedness of languages.
arXiv Detail & Related papers (2021-09-02T20:53:14Z) - Discovering Representation Sprachbund For Multilingual Pre-Training [139.05668687865688]
We generate language representation from multilingual pre-trained models and conduct linguistic analysis.
We cluster all the target languages into multiple groups and name each group as a representation sprachbund.
Experiments are conducted on cross-lingual benchmarks and significant improvements are achieved compared to strong baselines.
arXiv Detail & Related papers (2021-09-01T09:32:06Z) - Are Multilingual Models Effective in Code-Switching? [57.78477547424949]
We study the effectiveness of multilingual language models to understand their capability and adaptability to the mixed-language setting.
Our findings suggest that pre-trained multilingual models do not necessarily guarantee high-quality representations on code-switching.
arXiv Detail & Related papers (2021-03-24T16:20:02Z) - CoSDA-ML: Multi-Lingual Code-Switching Data Augmentation for Zero-Shot
Cross-Lingual NLP [68.2650714613869]
We propose a data augmentation framework to generate multi-lingual code-switching data to fine-tune mBERT.
Compared with the existing work, our method does not rely on bilingual sentences for training, and requires only one training process for multiple target languages.
arXiv Detail & Related papers (2020-06-11T13:15:59Z) - Finding Universal Grammatical Relations in Multilingual BERT [47.74015366712623]
We show that subspaces of mBERT representations recover syntactic tree distances in languages other than English.
We present an unsupervised analysis method that provides evidence mBERT learns representations of syntactic dependency labels.
arXiv Detail & Related papers (2020-05-09T20:46:02Z) - Learning to Scale Multilingual Representations for Vision-Language Tasks [51.27839182889422]
The effectiveness of SMALR is demonstrated with ten diverse languages, over twice the number supported in vision-language tasks to date.
We evaluate on multilingual image-sentence retrieval and outperform prior work by 3-4% with less than 1/5th the training parameters compared to other word embedding methods.
arXiv Detail & Related papers (2020-04-09T01:03:44Z) - XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating
Cross-lingual Generalization [128.37244072182506]
Cross-lingual TRansfer Evaluation of Multilinguals XTREME is a benchmark for evaluating the cross-lingual generalization capabilities of multilingual representations across 40 languages and 9 tasks.
We demonstrate that while models tested on English reach human performance on many tasks, there is still a sizable gap in the performance of cross-lingually transferred models.
arXiv Detail & Related papers (2020-03-24T19:09:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.