Zero-Shot Language Transfer vs Iterative Back Translation for
Unsupervised Machine Translation
- URL: http://arxiv.org/abs/2104.00106v1
- Date: Wed, 31 Mar 2021 20:47:19 GMT
- Title: Zero-Shot Language Transfer vs Iterative Back Translation for
Unsupervised Machine Translation
- Authors: Aviral Joshi, Chengzhi Huang, Har Simrat Singh
- Abstract summary: This work focuses on comparing different solutions for machine translation on low resource language pairs.
We discuss how the data size affects the performance of both unsupervised MT and transfer learning.
- Score: 1.2891210250935146
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work focuses on comparing different solutions for machine translation on
low resource language pairs, namely, with zero-shot transfer learning and
unsupervised machine translation. We discuss how the data size affects the
performance of both unsupervised MT and transfer learning. Additionally we also
look at how the domain of the data affects the result of unsupervised MT. The
code to all the experiments performed in this project are accessible on Github.
Related papers
- Towards Zero-Shot Multimodal Machine Translation [64.9141931372384]
We propose a method to bypass the need for fully supervised data to train multimodal machine translation systems.
Our method, called ZeroMMT, consists in adapting a strong text-only machine translation (MT) model by training it on a mixture of two objectives.
To prove that our method generalizes to languages with no fully supervised training data available, we extend the CoMMuTE evaluation dataset to three new languages: Arabic, Russian and Chinese.
arXiv Detail & Related papers (2024-07-18T15:20:31Z) - An approach for mistranslation removal from popular dataset for Indic MT
Task [5.4755933832880865]
We propose an algorithm to remove mistranslations from the training corpus and evaluate its performance and efficiency.
Two Indic languages (ILs), namely, Hindi (HIN) and Odia (ODI) are chosen for the experiment.
The quality of the translations in the experiment is evaluated using standard metrics such as BLEU, METEOR, and RIBES.
arXiv Detail & Related papers (2024-01-12T06:37:19Z) - On the Copying Problem of Unsupervised NMT: A Training Schedule with a
Language Discriminator Loss [120.19360680963152]
unsupervised neural machine translation (UNMT) has achieved success in many language pairs.
The copying problem, i.e., directly copying some parts of the input sentence as the translation, is common among distant language pairs.
We propose a simple but effective training schedule that incorporates a language discriminator loss.
arXiv Detail & Related papers (2023-05-26T18:14:23Z) - Improving Cascaded Unsupervised Speech Translation with Denoising
Back-translation [70.33052952571884]
We propose to build a cascaded speech translation system without leveraging any kind of paired data.
We use fully unpaired data to train our unsupervised systems and evaluate our results on CoVoST 2 and CVSS.
arXiv Detail & Related papers (2023-05-12T13:07:51Z) - Domain Mismatch Doesn't Always Prevent Cross-Lingual Transfer Learning [51.232774288403114]
Cross-lingual transfer learning has been surprisingly effective in zero-shot cross-lingual classification.
We show that a simple regimen can overcome much of the effect of domain mismatch in cross-lingual transfer.
arXiv Detail & Related papers (2022-11-30T01:24:33Z) - Original or Translated? A Causal Analysis of the Impact of
Translationese on Machine Translation Performance [31.47795931399995]
Human-translated text displays distinct features from naturally written text in the same language.
We find that existing work on translationese neglects some important factors and the conclusions are mostly correlational but not causal.
We show that these two factors have a large causal effect on the MT performance.
arXiv Detail & Related papers (2022-05-04T19:17:55Z) - It's Easier to Translate out of English than into it: Measuring Neural
Translation Difficulty by Cross-Mutual Information [90.35685796083563]
Cross-mutual information (XMI) is an asymmetric information-theoretic metric of machine translation difficulty.
XMI exploits the probabilistic nature of most neural machine translation models.
We present the first systematic and controlled study of cross-lingual translation difficulties using modern neural translation systems.
arXiv Detail & Related papers (2020-05-05T17:38:48Z) - When and Why is Unsupervised Neural Machine Translation Useless? [43.68079166777282]
In ten translation tasks with various data settings, we analyze the conditions under which the unsupervised methods fail to produce reasonable translations.
Our analyses pinpoint the limits of the current unsupervised NMT and also suggest immediate research directions.
arXiv Detail & Related papers (2020-04-22T14:00:55Z) - When Does Unsupervised Machine Translation Work? [23.690875724726908]
We conduct an empirical evaluation of unsupervised machine translation (MT) using dissimilar language pairs, dissimilar domains, diverse datasets, and authentic low-resource languages.
We find that performance rapidly deteriorates when source and target corpora are from different domains.
We additionally find that unsupervised MT performance declines when source and target languages use different scripts, and observe very poor performance on authentic low-resource language pairs.
arXiv Detail & Related papers (2020-04-12T00:57:47Z) - Cross-lingual Supervision Improves Unsupervised Neural Machine
Translation [97.84871088440102]
We introduce a multilingual unsupervised NMT framework to leverage weakly supervised signals from high-resource language pairs to zero-resource translation directions.
Method significantly improves the translation quality by more than 3 BLEU score on six benchmark unsupervised translation directions.
arXiv Detail & Related papers (2020-04-07T05:46:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.