Dynamic Context Selection for Document-level Neural Machine Translation
via Reinforcement Learning
- URL: http://arxiv.org/abs/2010.04314v1
- Date: Fri, 9 Oct 2020 01:05:32 GMT
- Title: Dynamic Context Selection for Document-level Neural Machine Translation
via Reinforcement Learning
- Authors: Xiaomian Kang, Yang Zhao, Jiajun Zhang, Chengqing Zong
- Abstract summary: We propose an effective approach to select dynamic context for document-level translation.
A novel reward is proposed to encourage the selection and utilization of dynamic context sentences.
Experiments demonstrate that our approach can select adaptive context sentences for different source sentences.
- Score: 55.18886832219127
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Document-level neural machine translation has yielded attractive
improvements. However, majority of existing methods roughly use all context
sentences in a fixed scope. They neglect the fact that different source
sentences need different sizes of context. To address this problem, we propose
an effective approach to select dynamic context so that the document-level
translation model can utilize the more useful selected context sentences to
produce better translations. Specifically, we introduce a selection module that
is independent of the translation module to score each candidate context
sentence. Then, we propose two strategies to explicitly select a variable
number of context sentences and feed them into the translation module. We train
the two modules end-to-end via reinforcement learning. A novel reward is
proposed to encourage the selection and utilization of dynamic context
sentences. Experiments demonstrate that our approach can select adaptive
context sentences for different source sentences, and significantly improves
the performance of document-level translation methods.
Related papers
- HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - Beyond Contrastive Learning: A Variational Generative Model for
Multilingual Retrieval [109.62363167257664]
We propose a generative model for learning multilingual text embeddings.
Our model operates on parallel data in $N$ languages.
We evaluate this method on a suite of tasks including semantic similarity, bitext mining, and cross-lingual question retrieval.
arXiv Detail & Related papers (2022-12-21T02:41:40Z) - SMDT: Selective Memory-Augmented Neural Document Translation [53.4627288890316]
We propose a Selective Memory-augmented Neural Document Translation model to deal with documents containing large hypothesis space of context.
We retrieve similar bilingual sentence pairs from the training corpus to augment global context.
We extend the two-stream attention model with selective mechanism to capture local context and diverse global contexts.
arXiv Detail & Related papers (2022-01-05T14:23:30Z) - Using Optimal Transport as Alignment Objective for fine-tuning
Multilingual Contextualized Embeddings [7.026476782041066]
We propose using Optimal Transport (OT) as an alignment objective during fine-tuning to improve multilingual contextualized representations.
This approach does not require word-alignment pairs prior to fine-tuning and instead learns the word alignments within context in an unsupervised manner.
arXiv Detail & Related papers (2021-10-06T16:13:45Z) - Contrastive Learning for Context-aware Neural Machine TranslationUsing
Coreference Information [14.671424999873812]
We propose CorefCL, a novel data augmentation and contrastive learning scheme based on coreference between the source and contextual sentences.
By corrupting automatically detected coreference mentions in the contextual sentence, CorefCL can train the model to be sensitive to coreference inconsistency.
In experiments, our method consistently improved BLEU of compared models on English-German and English-Korean tasks.
arXiv Detail & Related papers (2021-09-13T05:18:47Z) - Context-Adaptive Document-Level Neural Machine Translation [1.52292571922932]
We introduce a data-adaptive method that enables the model to adopt the necessary and useful context.
Experiments demonstrate the proposed approach can significantly improve the performance over the previous methods with a gain up to 1.99 BLEU points.
arXiv Detail & Related papers (2021-04-16T17:43:58Z) - Dynamic Data Selection and Weighting for Iterative Back-Translation [116.14378571769045]
We propose a curriculum learning strategy for iterative back-translation models.
We evaluate our models on domain adaptation, low-resource, and high-resource MT settings.
Experimental results demonstrate that our methods achieve improvements of up to 1.8 BLEU points over competitive baselines.
arXiv Detail & Related papers (2020-04-07T19:49:58Z) - Learning Contextualized Sentence Representations for Document-Level
Neural Machine Translation [59.191079800436114]
Document-level machine translation incorporates inter-sentential dependencies into the translation of a source sentence.
We propose a new framework to model cross-sentence dependencies by training neural machine translation (NMT) to predict both the target translation and surrounding sentences of a source sentence.
arXiv Detail & Related papers (2020-03-30T03:38:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.