Rapid Domain Adaptation for Machine Translation with Monolingual Data
- URL: http://arxiv.org/abs/2010.12652v1
- Date: Fri, 23 Oct 2020 20:31:37 GMT
- Title: Rapid Domain Adaptation for Machine Translation with Monolingual Data
- Authors: Mahdis Mahdieh, Mia Xu Chen, Yuan Cao, Orhan Firat
- Abstract summary: One challenge of machine translation is how to quickly adapt to unseen domains in face of surging events like COVID-19.
In this paper, we propose an approach that enables rapid domain adaptation from the perspective of unsupervised translation.
- Score: 31.70276147485463
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One challenge of machine translation is how to quickly adapt to unseen
domains in face of surging events like COVID-19, in which case timely and
accurate translation of in-domain information into multiple languages is
critical but little parallel data is available yet. In this paper, we propose
an approach that enables rapid domain adaptation from the perspective of
unsupervised translation. Our proposed approach only requires in-domain
monolingual data and can be quickly applied to a preexisting translation system
trained on general domain, reaching significant gains on in-domain translation
quality with little or no drop on general-domain. We also propose an effective
procedure of simultaneous adaptation for multiple domains and languages. To the
best of our knowledge, this is the first attempt that aims to address
unsupervised multilingual domain adaptation.
Related papers
- $m^4Adapter$: Multilingual Multi-Domain Adaptation for Machine
Translation with a Meta-Adapter [128.69723410769586]
Multilingual neural machine translation models (MNMT) yield state-of-the-art performance when evaluated on data from a domain and language pair.
When a MNMT model is used to translate under domain shift or to a new language pair, performance drops dramatically.
We propose $m4Adapter$, which combines domain and language knowledge using meta-learning with adapters.
arXiv Detail & Related papers (2022-10-21T12:25:05Z) - Non-Parametric Domain Adaptation for End-to-End Speech Translation [72.37869362559212]
End-to-End Speech Translation (E2E-ST) has received increasing attention due to the potential of its less error propagation, lower latency, and fewer parameters.
We propose a novel non-parametric method that leverages domain-specific text translation corpus to achieve domain adaptation for the E2E-ST system.
arXiv Detail & Related papers (2022-05-23T11:41:02Z) - Efficient Machine Translation Domain Adaptation [7.747003493657217]
Machine translation models struggle when translating out-of-domain text.
domain adaptation methods focus on fine-tuning or training the entire or part of the model on every new domain.
We introduce a simple but effective caching strategy that avoids performing retrieval when similar contexts have been seen before.
arXiv Detail & Related papers (2022-04-26T21:47:54Z) - Multilingual Domain Adaptation for NMT: Decoupling Language and Domain
Information with Adapters [66.7986513246294]
We study the compositionality of language and domain adapters in the context of Machine Translation.
We find that in the partial resource scenario a naive combination of domain-specific and language-specific adapters often results in catastrophic forgetting' of the missing languages.
arXiv Detail & Related papers (2021-10-18T18:55:23Z) - Non-Parametric Unsupervised Domain Adaptation for Neural Machine
Translation [61.27321597981737]
$k$NN-MT has shown the promising capability of directly incorporating the pre-trained neural machine translation (NMT) model with domain-specific token-level $k$-nearest-neighbor retrieval.
We propose a novel framework that directly uses in-domain monolingual sentences in the target language to construct an effective datastore for $k$-nearest-neighbor retrieval.
arXiv Detail & Related papers (2021-09-14T11:50:01Z) - Generalised Unsupervised Domain Adaptation of Neural Machine Translation
with Cross-Lingual Data Selection [34.90952499734384]
We propose a cross-lingual data selection method to extract in-domain sentences in the missing language side from a large generic monolingual corpus.
Our proposed method trains an adaptive layer on top of multilingual BERT by contrastive learning to align the representation between the source and target language.
We evaluate our cross-lingual data selection method on NMT across five diverse domains in three language pairs, as well as a real-world scenario of translation for COVID-19.
arXiv Detail & Related papers (2021-09-09T14:12:12Z) - Domain Adaptation and Multi-Domain Adaptation for Neural Machine
Translation: A Survey [9.645196221785694]
We focus on robust approaches to domain adaptation for Neural Machine Translation (NMT) models.
In particular, we look at the case where a system may need to translate sentences from multiple domains.
We highlight the benefits of domain adaptation and multi-domain adaptation techniques to other lines of NMT research.
arXiv Detail & Related papers (2021-04-14T16:21:37Z) - FDMT: A Benchmark Dataset for Fine-grained Domain Adaptation in Machine
Translation [53.87731008029645]
We present a real-world fine-grained domain adaptation task in machine translation (FDMT)
The FDMT dataset consists of four sub-domains of information technology: autonomous vehicles, AI education, real-time networks and smart phone.
We make quantitative experiments and deep analyses in this new setting, which benchmarks the fine-grained domain adaptation task.
arXiv Detail & Related papers (2020-12-31T17:15:09Z) - A Simple Baseline to Semi-Supervised Domain Adaptation for Machine
Translation [73.3550140511458]
State-of-the-art neural machine translation (NMT) systems are data-hungry and perform poorly on new domains with no supervised data.
We propose a simple but effect approach to the semi-supervised domain adaptation scenario of NMT.
This approach iteratively trains a Transformer-based NMT model via three training objectives: language modeling, back-translation, and supervised translation.
arXiv Detail & Related papers (2020-01-22T16:42:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.