A Comparison of Approaches to Document-level Machine Translation
- URL: http://arxiv.org/abs/2101.11040v1
- Date: Tue, 26 Jan 2021 19:21:09 GMT
- Title: A Comparison of Approaches to Document-level Machine Translation
- Authors: Zhiyi Ma, Sergey Edunov, Michael Auli
- Abstract summary: This paper presents a systematic comparison of selected approaches to document-level phenomena evaluation suites.
We find that a simple method based purely on back-translating monolingual document-level data performs as well as much more elaborate alternatives.
- Score: 34.2276281264886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Document-level machine translation conditions on surrounding sentences to
produce coherent translations. There has been much recent work in this area
with the introduction of custom model architectures and decoding algorithms.
This paper presents a systematic comparison of selected approaches from the
literature on two benchmarks for which document-level phenomena evaluation
suites exist. We find that a simple method based purely on back-translating
monolingual document-level data performs as well as much more elaborate
alternatives, both in terms of document-level metrics as well as human
evaluation.
Related papers
- Contextual Document Embeddings [77.22328616983417]
We propose two complementary methods for contextualized document embeddings.
First, an alternative contrastive learning objective that explicitly incorporates the document neighbors into the intra-batch contextual loss.
Second, a new contextual architecture that explicitly encodes neighbor document information into the encoded representation.
arXiv Detail & Related papers (2024-10-03T14:33:34Z) - End-to-End Evaluation for Low-Latency Simultaneous Speech Translation [55.525125193856084]
We propose the first framework to perform and evaluate the various aspects of low-latency speech translation under realistic conditions.
This includes the segmentation of the audio as well as the run-time of the different components.
We also compare different approaches to low-latency speech translation using this framework.
arXiv Detail & Related papers (2023-08-07T09:06:20Z) - On Search Strategies for Document-Level Neural Machine Translation [51.359400776242786]
Document-level neural machine translation (NMT) models produce a more consistent output across a document.
In this work, we aim to answer the question how to best utilize a context-aware translation model in decoding.
arXiv Detail & Related papers (2023-06-08T11:30:43Z) - HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - Discourse Cohesion Evaluation for Document-Level Neural Machine
Translation [36.96887050831173]
It is well known that translations generated by an excellent document-level neural machine translation (NMT) model are consistent and coherent.
Existing sentence-level evaluation metrics like BLEU can hardly reflect the model's performance at the document level.
We propose a new test suite that considers four cohesive manners to measure the cohesiveness of document translations.
arXiv Detail & Related papers (2022-08-19T01:56:00Z) - Document-level Neural Machine Translation with Document Embeddings [82.4684444847092]
This work focuses on exploiting detailed document-level context in terms of multiple forms of document embeddings.
The proposed document-aware NMT is implemented to enhance the Transformer baseline by introducing both global and local document-level clues on the source end.
arXiv Detail & Related papers (2020-09-16T19:43:29Z) - Towards Making the Most of Context in Neural Machine Translation [112.9845226123306]
We argue that previous research did not make a clear use of the global context.
We propose a new document-level NMT framework that deliberately models the local context of each sentence.
arXiv Detail & Related papers (2020-02-19T03:30:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.