Unsupervised Transfer Learning in Multilingual Neural Machine
Translation with Cross-Lingual Word Embeddings
- URL: http://arxiv.org/abs/2103.06689v1
- Date: Thu, 11 Mar 2021 14:22:08 GMT
- Title: Unsupervised Transfer Learning in Multilingual Neural Machine
Translation with Cross-Lingual Word Embeddings
- Authors: Carlos Mullov and Ngoc-Quan Pham and Alexander Waibel
- Abstract summary: We exploit a language independent multilingual sentence representation to easily generalize to a new language.
Blindly decoding from Portuguese using a basesystem containing several Romance languages we achieve scores of 36.4 BLEU for Portuguese-English and 12.8 BLEU for Russian-English.
We explore a more practical adaptation approach through non-iterative backtranslation, exploiting our model's ability to produce high quality translations.
- Score: 72.69253034282035
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we look into adding a new language to a multilingual NMT system
in an unsupervised fashion. Under the utilization of pre-trained cross-lingual
word embeddings we seek to exploit a language independent multilingual sentence
representation to easily generalize to a new language. While using
cross-lingual embeddings for word lookup we decode from a yet entirely unseen
source language in a process we call blind decoding. Blindly decoding from
Portuguese using a basesystem containing several Romance languages we achieve
scores of 36.4 BLEU for Portuguese-English and 12.8 BLEU for Russian-English.
In an attempt to train the mapping from the encoder sentence representation to
a new target language we use our model as an autoencoder. Merely training to
translate from Portuguese to Portuguese while freezing the encoder we achieve
26 BLEU on English-Portuguese, and up to 28 BLEU when adding artificial noise
to the input. Lastly we explore a more practical adaptation approach through
non-iterative backtranslation, exploiting our model's ability to produce high
quality translations through blind decoding. This yields us up to 34.6 BLEU on
English-Portuguese, attaining near parity with a model adapted on real
bilingual data.
Related papers
- Decoupled Vocabulary Learning Enables Zero-Shot Translation from Unseen Languages [55.157295899188476]
neural machine translation systems learn to map sentences of different languages into a common representation space.
In this work, we test this hypothesis by zero-shot translating from unseen languages.
We demonstrate that this setup enables zero-shot translation from entirely unseen languages.
arXiv Detail & Related papers (2024-08-05T07:58:58Z) - CoVoSwitch: Machine Translation of Synthetic Code-Switched Text Based on Intonation Units [0.0]
We synthesize code-switching data by replacing intonation units detected through PSST.
We evaluate the code-switching translation performance of two multilingual translation models, M2M-100 418M and NLLB-200 600M.
arXiv Detail & Related papers (2024-07-19T13:26:35Z) - On the Off-Target Problem of Zero-Shot Multilingual Neural Machine
Translation [104.85258654917297]
We find that failing in encoding discriminative target language signal will lead to off-target and a closer lexical distance.
We propose Language Aware Vocabulary Sharing (LAVS) to construct the multilingual vocabulary.
We conduct experiments on a multilingual machine translation benchmark in 11 languages.
arXiv Detail & Related papers (2023-05-18T12:43:31Z) - Chain-of-Dictionary Prompting Elicits Translation in Large Language Models [100.47154959254937]
Large language models (LLMs) have shown surprisingly good performance in multilingual neural machine translation (MNMT)
We present a novel method, CoD, which augments LLMs with prior knowledge with the chains of multilingual dictionaries for a subset of input words to elicit translation abilities.
arXiv Detail & Related papers (2023-05-11T05:19:47Z) - Exposing Cross-Lingual Lexical Knowledge from Multilingual Sentence
Encoders [85.80950708769923]
We probe multilingual language models for the amount of cross-lingual lexical knowledge stored in their parameters, and compare them against the original multilingual LMs.
We also devise a novel method to expose this knowledge by additionally fine-tuning multilingual models.
We report substantial gains on standard benchmarks.
arXiv Detail & Related papers (2022-04-30T13:23:16Z) - Call Larisa Ivanovna: Code-Switching Fools Multilingual NLU Models [1.827510863075184]
Novel benchmarks for multilingual natural language understanding (NLU) include monolingual sentences in several languages, annotated with intents and slots.
Existing benchmarks lack of code-switched utterances, which are difficult to gather and label due to complexity in the grammatical structure.
Our work adopts recognized methods to generate plausible and naturally-sounding code-switched utterances and uses them to create a synthetic code-switched test set.
arXiv Detail & Related papers (2021-09-29T11:15:00Z) - Learning Contextualised Cross-lingual Word Embeddings and Alignments for
Extremely Low-Resource Languages Using Parallel Corpora [63.5286019659504]
We propose a new approach for learning contextualised cross-lingual word embeddings based on a small parallel corpus.
Our method obtains word embeddings via an LSTM encoder-decoder model that simultaneously translates and reconstructs an input sentence.
arXiv Detail & Related papers (2020-10-27T22:24:01Z) - HausaMT v1.0: Towards English-Hausa Neural Machine Translation [0.012691047660244334]
We build a baseline model for English-Hausa machine translation.
The Hausa language is the second largest Afro-Asiatic language in the world after Arabic.
arXiv Detail & Related papers (2020-06-09T02:08:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.