Re-translation versus Streaming for Simultaneous Translation
- URL: http://arxiv.org/abs/2004.03643v3
- Date: Mon, 29 Jun 2020 23:36:13 GMT
- Title: Re-translation versus Streaming for Simultaneous Translation
- Authors: Naveen Arivazhagan, Colin Cherry, Wolfgang Macherey and George Foster
- Abstract summary: We study a problem in which revisions to the hypothesis beyond strictly appending words are permitted.
In this setting, we compare custom streaming approaches to re-translation.
We find re-translation to be as good or better than state-of-the-art streaming systems.
- Score: 14.800214853561823
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There has been great progress in improving streaming machine translation, a
simultaneous paradigm where the system appends to a growing hypothesis as more
source content becomes available. We study a related problem in which revisions
to the hypothesis beyond strictly appending words are permitted. This is
suitable for applications such as live captioning an audio feed. In this
setting, we compare custom streaming approaches to re-translation, a
straightforward strategy where each new source token triggers a distinct
translation from scratch. We find re-translation to be as good or better than
state-of-the-art streaming systems, even when operating under constraints that
allow very few revisions. We attribute much of this success to a previously
proposed data-augmentation technique that adds prefix-pairs to the training
data, which alongside wait-k inference forms a strong baseline for streaming
translation. We also highlight re-translation's ability to wrap arbitrarily
powerful MT systems with an experiment showing large improvements from an
upgrade to its base model.
Related papers
- Contextual Refinement of Translations: Large Language Models for Sentence and Document-Level Post-Editing [12.843274390224853]
Large Language Models (LLM's) have demonstrated considerable success in various Natural Language Processing tasks.
We show that they have yet to attain state-of-the-art performance in Neural Machine Translation.
We propose adapting LLM's as Automatic Post-Editors (APE) rather than direct translators.
arXiv Detail & Related papers (2023-10-23T12:22:15Z) - DiariST: Streaming Speech Translation with Speaker Diarization [53.595990270899414]
We propose DiariST, the first streaming ST and SD solution.
It is built upon a neural transducer-based streaming ST system and integrates token-level serialized output training and t-vector.
Our system achieves a strong ST and SD capability compared to offline systems based on Whisper, while performing streaming inference for overlapping speech.
arXiv Detail & Related papers (2023-09-14T19:33:27Z) - Principled Paraphrase Generation with Parallel Corpora [52.78059089341062]
We formalize the implicit similarity function induced by round-trip Machine Translation.
We show that it is susceptible to non-paraphrase pairs sharing a single ambiguous translation.
We design an alternative similarity metric that mitigates this issue.
arXiv Detail & Related papers (2022-05-24T17:22:42Z) - From Simultaneous to Streaming Machine Translation by Leveraging
Streaming History [4.831134508326648]
Simultaneous Machine Translation is the task of incrementally translating an input sentence before it is fully available.
Streaming MT can be understood as an extension of Simultaneous MT to the incremental translation of a continuous input text stream.
In this work, a state-of-the-art simultaneous sentence-level MT system is extended to the streaming setup by leveraging the streaming history.
arXiv Detail & Related papers (2022-03-04T17:41:45Z) - Anticipation-free Training for Simultaneous Translation [70.85761141178597]
Simultaneous translation (SimulMT) speeds up the translation process by starting to translate before the source sentence is completely available.
Existing methods increase latency or introduce adaptive read-write policies for SimulMT models to handle local reordering and improve translation quality.
We propose a new framework that decomposes the translation process into the monotonic translation step and the reordering step.
arXiv Detail & Related papers (2022-01-30T16:29:37Z) - Faster Re-translation Using Non-Autoregressive Model For Simultaneous
Neural Machine Translation [10.773010211146694]
We propose a faster re-translation system based on a non-autoregressive sequence generation model (FReTNA)
The proposed model reduces the average computation time by a factor of 20 when compared to the ReTA model.
It also outperforms the streaming-based Wait-k model both in terms of time (1.5 times lower) and translation quality.
arXiv Detail & Related papers (2020-12-29T09:43:27Z) - SimulEval: An Evaluation Toolkit for Simultaneous Translation [59.02724214432792]
Simultaneous translation on both text and speech focuses on a real-time and low-latency scenario.
SimulEval is an easy-to-use and general evaluation toolkit for both simultaneous text and speech translation.
arXiv Detail & Related papers (2020-07-31T17:44:41Z) - Self-Supervised Representations Improve End-to-End Speech Translation [57.641761472372814]
We show that self-supervised pre-trained features can consistently improve the translation performance.
Cross-lingual transfer allows to extend to a variety of languages without or with little tuning.
arXiv Detail & Related papers (2020-06-22T10:28:38Z) - Simplify-then-Translate: Automatic Preprocessing for Black-Box Machine
Translation [5.480070710278571]
We introduce a method to improve black-box machine translation systems via automatic pre-processing (APP) using sentence simplification.
We first propose a method to automatically generate a large in-domain paraphrase corpus through back-translation with a black-box MT system.
We show that this preprocessing leads to better translation performance as compared to non-preprocessed source sentences.
arXiv Detail & Related papers (2020-05-22T14:15:53Z) - Learning Contextualized Sentence Representations for Document-Level
Neural Machine Translation [59.191079800436114]
Document-level machine translation incorporates inter-sentential dependencies into the translation of a source sentence.
We propose a new framework to model cross-sentence dependencies by training neural machine translation (NMT) to predict both the target translation and surrounding sentences of a source sentence.
arXiv Detail & Related papers (2020-03-30T03:38:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.