Contextual Neural Machine Translation Improves Translation of Cataphoric
Pronouns
- URL: http://arxiv.org/abs/2004.09894v2
- Date: Tue, 28 Apr 2020 08:27:57 GMT
- Title: Contextual Neural Machine Translation Improves Translation of Cataphoric
Pronouns
- Authors: KayYen Wong, Sameen Maruf, Gholamreza Haffari
- Abstract summary: We investigate the effect of future sentences as context by comparing the performance of a contextual NMT model trained with the future context to the one trained with the past context.
Our experiments and evaluation, using generic and pronoun-focused automatic metrics, show that the use of future context achieves significant improvements over the context-agnostic Transformer.
- Score: 50.245845110446496
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The advent of context-aware NMT has resulted in promising improvements in the
overall translation quality and specifically in the translation of discourse
phenomena such as pronouns. Previous works have mainly focused on the use of
past sentences as context with a focus on anaphora translation. In this work,
we investigate the effect of future sentences as context by comparing the
performance of a contextual NMT model trained with the future context to the
one trained with the past context. Our experiments and evaluation, using
generic and pronoun-focused automatic metrics, show that the use of future
context not only achieves significant improvements over the context-agnostic
Transformer, but also demonstrates comparable and in some cases improved
performance over its counterpart trained on past context. We also perform an
evaluation on a targeted cataphora test suite and report significant gains over
the context-agnostic Transformer in terms of BLEU.
Related papers
- A Case Study on Contextual Machine Translation in a Professional Scenario of Subtitling [3.925328332747599]
We report on an industrial case study carried out to investigate the benefit of machine translation (MT) in a professional scenario of translating TV subtitles.
We found that post-editors marked significantly fewer context-related errors when correcting the outputs of MTCue, the context-aware model.
We also present the results of a survey of the employed post-editors, which highlights contextual inadequacy as a significant gap consistently observed in MT.
arXiv Detail & Related papers (2024-06-27T11:20:14Z) - Understanding and Addressing the Under-Translation Problem from the Perspective of Decoding Objective [72.83966378613238]
Under-translation and over-translation remain two challenging problems in state-of-the-art Neural Machine Translation (NMT) systems.
We conduct an in-depth analysis on the underlying cause of under-translation in NMT, providing an explanation from the perspective of decoding objective.
We propose employing the confidence of predicting End Of Sentence (EOS) as a detector for under-translation, and strengthening the confidence-based penalty to penalize candidates with a high risk of under-translation.
arXiv Detail & Related papers (2024-05-29T09:25:49Z) - Context-aware Neural Machine Translation for English-Japanese Business
Scene Dialogues [14.043741721036543]
This paper explores how context-awareness can improve the performance of the current Neural Machine Translation (NMT) models for English-Japanese business dialogues translation.
We propose novel context tokens encoding extra-sentential information, such as speaker turn and scene type.
We find that models leverage both preceding sentences and extra-sentential context (with CXMI increasing with context size) and we provide a more focused analysis on honorifics translation.
arXiv Detail & Related papers (2023-11-20T18:06:03Z) - Challenges in Context-Aware Neural Machine Translation [39.89082986080746]
Context-aware neural machine translation involves leveraging information beyond sentence-level context to resolve discourse dependencies.
Despite well-reasoned intuitions, most context-aware translation models show only modest improvements over sentence-level systems.
We propose a more realistic setting for document-level translation, called paragraph-to-paragraph (para2para) translation.
arXiv Detail & Related papers (2023-05-23T07:08:18Z) - HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - When Does Translation Require Context? A Data-driven, Multilingual
Exploration [71.43817945875433]
proper handling of discourse significantly contributes to the quality of machine translation (MT)
Recent works in context-aware MT attempt to target a small set of discourse phenomena during evaluation.
We develop the Multilingual Discourse-Aware benchmark, a series of taggers that identify and evaluate model performance on discourse phenomena.
arXiv Detail & Related papers (2021-09-15T17:29:30Z) - Measuring and Increasing Context Usage in Context-Aware Machine
Translation [64.5726087590283]
We introduce a new metric, conditional cross-mutual information, to quantify the usage of context by machine translation models.
We then introduce a new, simple training method, context-aware word dropout, to increase the usage of context by context-aware models.
arXiv Detail & Related papers (2021-05-07T19:55:35Z) - Can Your Context-Aware MT System Pass the DiP Benchmark Tests? :
Evaluation Benchmarks for Discourse Phenomena in Machine Translation [7.993547048820065]
We introduce the first of their kind MT benchmark datasets that aim to track and hail improvements across four main discourse phenomena.
Surprisingly, we find that existing context-aware models do not improve discourse-related translations consistently across languages and phenomena.
arXiv Detail & Related papers (2020-04-30T07:15:36Z) - Multilingual Alignment of Contextual Word Representations [49.42244463346612]
BERT exhibits significantly improved zero-shot performance on XNLI compared to the base model.
We introduce a contextual version of word retrieval and show that it correlates well with downstream zero-shot transfer.
These results support contextual alignment as a useful concept for understanding large multilingual pre-trained models.
arXiv Detail & Related papers (2020-02-10T03:27:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.