Machine-Created Universal Language for Cross-lingual Transfer
- URL: http://arxiv.org/abs/2305.13071v2
- Date: Sun, 17 Dec 2023 03:20:13 GMT
- Title: Machine-Created Universal Language for Cross-lingual Transfer
- Authors: Yaobo Liang, Quanzhi Zhu, Junhe Zhao and Nan Duan
- Abstract summary: We propose a new Machine-created Universal Language (MUL) as an alternative intermediate language.
MUL comprises a set of discrete symbols forming a universal vocabulary and a natural language to MUL translator.
MUL unifies shared concepts from various languages into a single universal word, enhancing cross-language transfer.
- Score: 73.44138687502294
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There are two primary approaches to addressing cross-lingual transfer:
multilingual pre-training, which implicitly aligns the hidden representations
of various languages, and translate-test, which explicitly translates different
languages into an intermediate language, such as English. Translate-test offers
better interpretability compared to multilingual pre-training. However, it has
lower performance than multilingual pre-training(Conneau and Lample, 2019;
Conneau et al, 2020) and struggles with word-level tasks due to translation
altering word order. As a result, we propose a new Machine-created Universal
Language (MUL) as an alternative intermediate language. MUL comprises a set of
discrete symbols forming a universal vocabulary and a natural language to MUL
translator for converting multiple natural languages to MUL. MUL unifies shared
concepts from various languages into a single universal word, enhancing
cross-language transfer. Additionally, MUL retains language-specific words and
word order, allowing the model to be easily applied to word-level tasks. Our
experiments demonstrate that translating into MUL yields improved performance
compared to multilingual pre-training, and our analysis indicates that MUL
possesses strong interpretability. The code is at:
https://github.com/microsoft/Unicoder/tree/master/MCUL.
Related papers
- m3P: Towards Multimodal Multilingual Translation with Multimodal Prompt [39.2728779674405]
We propose a framework to leverage the multimodal prompt to guide the Multimodal Multilingual neural Machine Translation (m3P)
Our method aims to minimize the representation distance of different languages by regarding the image as a central language.
Experimental results show that m3P outperforms previous text-only baselines and multilingual multimodal methods by a large margin.
arXiv Detail & Related papers (2024-03-26T10:04:24Z) - Romanization-based Large-scale Adaptation of Multilingual Language
Models [124.57923286144515]
Large multilingual pretrained language models (mPLMs) have become the de facto state of the art for cross-lingual transfer in NLP.
We study and compare a plethora of data- and parameter-efficient strategies for adapting the mPLMs to romanized and non-romanized corpora of 14 diverse low-resource languages.
Our results reveal that UROMAN-based transliteration can offer strong performance for many languages, with particular gains achieved in the most challenging setups.
arXiv Detail & Related papers (2023-04-18T09:58:34Z) - LVP-M3: Language-aware Visual Prompt for Multilingual Multimodal Machine
Translation [94.33019040320507]
Multimodal Machine Translation (MMT) focuses on enhancing text-only translation with visual features.
Recent advances still struggle to train a separate model for each language pair, which is costly and unaffordable when the number of languages increases.
We propose the Multilingual MMT task by establishing two new Multilingual MMT benchmark datasets covering seven languages.
arXiv Detail & Related papers (2022-10-19T12:21:39Z) - Syntax-augmented Multilingual BERT for Cross-lingual Transfer [37.99210035238424]
This work shows that explicitly providing language syntax and training mBERT helps cross-lingual transfer.
Experiment results show that syntax-augmented mBERT improves cross-lingual transfer on popular benchmarks.
arXiv Detail & Related papers (2021-06-03T21:12:50Z) - VECO: Variable and Flexible Cross-lingual Pre-training for Language
Understanding and Generation [77.82373082024934]
We plug a cross-attention module into the Transformer encoder to explicitly build the interdependence between languages.
It can effectively avoid the degeneration of predicting masked words only conditioned on the context in its own language.
The proposed cross-lingual model delivers new state-of-the-art results on various cross-lingual understanding tasks of the XTREME benchmark.
arXiv Detail & Related papers (2020-10-30T03:41:38Z) - FILTER: An Enhanced Fusion Method for Cross-lingual Language
Understanding [85.29270319872597]
We propose an enhanced fusion method that takes cross-lingual data as input for XLM finetuning.
During inference, the model makes predictions based on the text input in the target language and its translation in the source language.
To tackle this issue, we propose an additional KL-divergence self-teaching loss for model training, based on auto-generated soft pseudo-labels for translated text in the target language.
arXiv Detail & Related papers (2020-09-10T22:42:15Z) - Learning to Scale Multilingual Representations for Vision-Language Tasks [51.27839182889422]
The effectiveness of SMALR is demonstrated with ten diverse languages, over twice the number supported in vision-language tasks to date.
We evaluate on multilingual image-sentence retrieval and outperform prior work by 3-4% with less than 1/5th the training parameters compared to other word embedding methods.
arXiv Detail & Related papers (2020-04-09T01:03:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.