Simultaneous Multi-Pivot Neural Machine Translation
- URL: http://arxiv.org/abs/2104.07410v1
- Date: Thu, 15 Apr 2021 12:19:52 GMT
- Title: Simultaneous Multi-Pivot Neural Machine Translation
- Authors: Raj Dabre, Aizhan Imankulova, Masahiro Kaneko, Abhisek Chakrabarty
- Abstract summary: In a simultaneous pivot NMT setting, using two pivot languages can lead to an improvement of up to 5.8 BLEU.
Our experiments in a low-resource setting using the N-way parallel UN corpus for Arabic to English NMT via French and Spanish as pivots reveals that in a simultaneous pivot NMT setting, using two pivot languages can lead to an improvement of up to 5.8 BLEU.
- Score: 12.796775798210133
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Parallel corpora are indispensable for training neural machine translation
(NMT) models, and parallel corpora for most language pairs do not exist or are
scarce. In such cases, pivot language NMT can be helpful where a pivot language
is used such that there exist parallel corpora between the source and pivot and
pivot and target languages. Naturally, the quality of pivot language
translation is more inferior to what could be achieved with a direct parallel
corpus of a reasonable size for that pair. In a real-time simultaneous
translation setting, the quality of pivot language translation deteriorates
even further given that the model has to output translations the moment a few
source words become available. To solve this issue, we propose multi-pivot
translation and apply it to a simultaneous translation setting involving pivot
languages. Our approach involves simultaneously translating a source language
into multiple pivots, which are then simultaneously translated together into
the target language by leveraging multi-source NMT. Our experiments in a
low-resource setting using the N-way parallel UN corpus for Arabic to English
NMT via French and Spanish as pivots reveals that in a simultaneous pivot NMT
setting, using two pivot languages can lead to an improvement of up to 5.8
BLEU.
Related papers
- How effective is Multi-source pivoting for Translation of Low Resource Indian Languages? [43.44411629370054]
This paper explores the'multi-source translation' approach with pivoting, using both source and pivot sentences to improve translation.
We find that multi-source pivoting yields marginal improvements over the state-of-the-art, contrary to previous claims.
arXiv Detail & Related papers (2024-06-19T08:31:52Z) - Extending Multilingual Machine Translation through Imitation Learning [60.15671816513614]
Imit-MNMT treats the task as an imitation learning process, which mimicks the behavior of an expert.
We show that our approach significantly improves the translation performance between the new and the original languages.
We also demonstrate that our approach is capable of solving copy and off-target problems.
arXiv Detail & Related papers (2023-11-14T21:04:03Z) - Direct Neural Machine Translation with Task-level Mixture of Experts models [1.2338729811609357]
Direct neural machine translation (direct NMT) translates text between two non-English languages.
Task-level Mixture of expert models (Task-level MoE) has shown promising NMT performance for a large number of language pairs.
arXiv Detail & Related papers (2023-10-18T18:19:45Z) - Mitigating Data Imbalance and Representation Degeneration in
Multilingual Machine Translation [103.90963418039473]
Bi-ACL is a framework that uses only target-side monolingual data and a bilingual dictionary to improve the performance of the MNMT model.
We show that Bi-ACL is more effective both in long-tail languages and in high-resource languages.
arXiv Detail & Related papers (2023-05-22T07:31:08Z) - Language Agnostic Multilingual Information Retrieval with Contrastive
Learning [59.26316111760971]
We present an effective method to train multilingual information retrieval systems.
We leverage parallel and non-parallel corpora to improve the pretrained multilingual language models.
Our model can work well even with a small number of parallel sentences.
arXiv Detail & Related papers (2022-10-12T23:53:50Z) - Multilingual Bidirectional Unsupervised Translation Through Multilingual
Finetuning and Back-Translation [23.401781865904386]
We propose a two-stage approach for training a single NMT model to translate unseen languages both to and from English.
For the first stage, we initialize an encoder-decoder model to pretrained XLM-R and RoBERTa weights, then perform multilingual fine-tuning on parallel data in 40 languages to English.
For the second stage, we leverage this generalization ability to generate synthetic parallel data from monolingual datasets, then bidirectionally train with successive rounds of back-translation.
arXiv Detail & Related papers (2022-09-06T21:20:41Z) - Towards the Next 1000 Languages in Multilingual Machine Translation:
Exploring the Synergy Between Supervised and Self-Supervised Learning [48.15259834021655]
We present a pragmatic approach towards building a multilingual machine translation model that covers hundreds of languages.
We use a mixture of supervised and self-supervised objectives, depending on the data availability for different language pairs.
We demonstrate that the synergy between these two training paradigms enables the model to produce high-quality translations in the zero-resource setting.
arXiv Detail & Related papers (2022-01-09T23:36:44Z) - Towards Reinforcement Learning for Pivot-based Neural Machine
Translation with Non-autoregressive Transformer [49.897891031932545]
Pivot-based neural machine translation (NMT) is commonly used in low-resource setups.
We present an end-to-end pivot-based integrated model, enabling training on source-target data.
arXiv Detail & Related papers (2021-09-27T14:49:35Z) - Pre-training Multilingual Neural Machine Translation by Leveraging
Alignment Information [72.2412707779571]
mRASP is an approach to pre-train a universal multilingual neural machine translation model.
We carry out experiments on 42 translation directions across a diverse setting, including low, medium, rich resource, and as well as transferring to exotic language pairs.
arXiv Detail & Related papers (2020-10-07T03:57:54Z) - Improving Massively Multilingual Neural Machine Translation and
Zero-Shot Translation [81.7786241489002]
Massively multilingual models for neural machine translation (NMT) are theoretically attractive, but often underperform bilingual models and deliver poor zero-shot translations.
We argue that multilingual NMT requires stronger modeling capacity to support language pairs with varying typological characteristics.
We propose random online backtranslation to enforce the translation of unseen training language pairs.
arXiv Detail & Related papers (2020-04-24T17:21:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.