A Simple Baseline to Semi-Supervised Domain Adaptation for Machine
Translation
- URL: http://arxiv.org/abs/2001.08140v2
- Date: Sat, 6 Jun 2020 02:45:10 GMT
- Title: A Simple Baseline to Semi-Supervised Domain Adaptation for Machine
Translation
- Authors: Di Jin, Zhijing Jin, Joey Tianyi Zhou, Peter Szolovits
- Abstract summary: State-of-the-art neural machine translation (NMT) systems are data-hungry and perform poorly on new domains with no supervised data.
We propose a simple but effect approach to the semi-supervised domain adaptation scenario of NMT.
This approach iteratively trains a Transformer-based NMT model via three training objectives: language modeling, back-translation, and supervised translation.
- Score: 73.3550140511458
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State-of-the-art neural machine translation (NMT) systems are data-hungry and
perform poorly on new domains with no supervised data. As data collection is
expensive and infeasible in many cases, domain adaptation methods are needed.
In this work, we propose a simple but effect approach to the semi-supervised
domain adaptation scenario of NMT, where the aim is to improve the performance
of a translation model on the target domain consisting of only non-parallel
data with the help of supervised source domain data. This approach iteratively
trains a Transformer-based NMT model via three training objectives: language
modeling, back-translation, and supervised translation. We evaluate this method
on two adaptation settings: adaptation between specific domains and adaptation
from a general domain to specific domains, and on two language pairs: German to
English and Romanian to English. With substantial performance improvement
achieved---up to +19.31 BLEU over the strongest baseline, and +47.69 BLEU
improvement over the unadapted model---we present this method as a simple but
tough-to-beat baseline in the field of semi-supervised domain adaptation for
NMT.
Related papers
- Non-Parametric Domain Adaptation for End-to-End Speech Translation [72.37869362559212]
End-to-End Speech Translation (E2E-ST) has received increasing attention due to the potential of its less error propagation, lower latency, and fewer parameters.
We propose a novel non-parametric method that leverages domain-specific text translation corpus to achieve domain adaptation for the E2E-ST system.
arXiv Detail & Related papers (2022-05-23T11:41:02Z) - Non-Parametric Unsupervised Domain Adaptation for Neural Machine
Translation [61.27321597981737]
$k$NN-MT has shown the promising capability of directly incorporating the pre-trained neural machine translation (NMT) model with domain-specific token-level $k$-nearest-neighbor retrieval.
We propose a novel framework that directly uses in-domain monolingual sentences in the target language to construct an effective datastore for $k$-nearest-neighbor retrieval.
arXiv Detail & Related papers (2021-09-14T11:50:01Z) - Domain Adaptation and Multi-Domain Adaptation for Neural Machine
Translation: A Survey [9.645196221785694]
We focus on robust approaches to domain adaptation for Neural Machine Translation (NMT) models.
In particular, we look at the case where a system may need to translate sentences from multiple domains.
We highlight the benefits of domain adaptation and multi-domain adaptation techniques to other lines of NMT research.
arXiv Detail & Related papers (2021-04-14T16:21:37Z) - Iterative Domain-Repaired Back-Translation [50.32925322697343]
In this paper, we focus on the domain-specific translation with low resources, where in-domain parallel corpora are scarce or nonexistent.
We propose a novel iterative domain-repaired back-translation framework, which introduces the Domain-Repair model to refine translations in synthetic bilingual data.
Experiments on adapting NMT models between specific domains and from the general domain to specific domains demonstrate the effectiveness of our proposed approach.
arXiv Detail & Related papers (2020-10-06T04:38:09Z) - Vocabulary Adaptation for Distant Domain Adaptation in Neural Machine
Translation [14.390932594872233]
Domain adaptation between distant domains cannot be performed effectively due to mismatches in vocabulary.
We propose vocabulary adaptation, a simple method for effective fine-tuning.
Our method improves the performance of conventional fine-tuning by 3.86 and 3.28 BLEU points in En-Ja and De-En translation.
arXiv Detail & Related papers (2020-04-30T14:27:59Z) - Dynamic Data Selection and Weighting for Iterative Back-Translation [116.14378571769045]
We propose a curriculum learning strategy for iterative back-translation models.
We evaluate our models on domain adaptation, low-resource, and high-resource MT settings.
Experimental results demonstrate that our methods achieve improvements of up to 1.8 BLEU points over competitive baselines.
arXiv Detail & Related papers (2020-04-07T19:49:58Z) - Supervised Domain Adaptation using Graph Embedding [86.3361797111839]
Domain adaptation methods assume that distributions between the two domains are shifted and attempt to realign them.
We propose a generic framework based on graph embedding.
We show that the proposed approach leads to a powerful Domain Adaptation framework.
arXiv Detail & Related papers (2020-03-09T12:25:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.