LEAPT: Learning Adaptive Prefix-to-prefix Translation For Simultaneous
Machine Translation
- URL: http://arxiv.org/abs/2303.11750v1
- Date: Tue, 21 Mar 2023 11:17:37 GMT
- Title: LEAPT: Learning Adaptive Prefix-to-prefix Translation For Simultaneous
Machine Translation
- Authors: Lei Lin, Shuangtao Li, Xiaodong Shi
- Abstract summary: Simultaneous machine translation is useful in many live scenarios but very challenging due to the trade-off between accuracy and latency.
We propose a novel adaptive training policy called LEAPT, which allows our machine translation model to learn how to translate source prefixes and make use of the future context.
- Score: 6.411228564798412
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Simultaneous machine translation, which aims at a real-time translation, is
useful in many live scenarios but very challenging due to the trade-off between
accuracy and latency. To achieve the balance for both, the model needs to wait
for appropriate streaming text (READ policy) and then generates its translation
(WRITE policy). However, WRITE policies of previous work either are specific to
the method itself due to the end-to-end training or suffer from the input
mismatch between training and decoding for the non-end-to-end training.
Therefore, it is essential to learn a generic and better WRITE policy for
simultaneous machine translation. Inspired by strategies utilized by human
interpreters and "wait" policies, we propose a novel adaptive prefix-to-prefix
training policy called LEAPT, which allows our machine translation model to
learn how to translate source sentence prefixes and make use of the future
context. Experiments show that our proposed methods greatly outperform
competitive baselines and achieve promising results.
Related papers
- PsFuture: A Pseudo-Future-based Zero-Shot Adaptive Policy for Simultaneous Machine Translation [8.1299957975257]
Simultaneous Machine Translation (SiMT) requires target tokens to be generated in real-time as streaming source tokens are consumed.
We propose PsFuture, the first zero-shot adaptive read/write policy for SiMT.
We introduce a novel training strategy, Prefix-to-Full (P2F), specifically tailored to adjust offline translation models for SiMT applications.
arXiv Detail & Related papers (2024-10-05T08:06:33Z) - Language Model is a Branch Predictor for Simultaneous Machine
Translation [73.82754138171587]
We propose incorporating branch prediction techniques in SiMT tasks to reduce translation latency.
We utilize a language model as a branch predictor to predict potential branch directions.
When the actual source word deviates from the predicted source word, we use the real source word to decode the output again, replacing the predicted output.
arXiv Detail & Related papers (2023-12-22T07:32:47Z) - Glancing Future for Simultaneous Machine Translation [35.46823126036308]
We propose a novel method to bridge the gap between the prefix2 training and seq2seq training.
We gradually reduce the available source information from the whole sentence to the prefix corresponding to that latency.
Our method is applicable to a wide range of SiMT methods and experiments demonstrate that our method outperforms strong baselines.
arXiv Detail & Related papers (2023-09-12T12:46:20Z) - Learning Optimal Policy for Simultaneous Machine Translation via Binary
Search [17.802607889752736]
Simultaneous machine translation (SiMT) starts to output translation while reading the source sentence.
The policy determines the number of source tokens read during the translation of each target token.
We present a new method for constructing the optimal policy online via binary search.
arXiv Detail & Related papers (2023-05-22T07:03:06Z) - M-Adapter: Modality Adaptation for End-to-End Speech-to-Text Translation [66.92823764664206]
We propose M-Adapter, a novel Transformer-based module, to adapt speech representations to text.
While shrinking the speech sequence, M-Adapter produces features desired for speech-to-text translation.
Our experimental results show that our model outperforms a strong baseline by up to 1 BLEU.
arXiv Detail & Related papers (2022-07-03T04:26:53Z) - Data-Driven Adaptive Simultaneous Machine Translation [51.01779863078624]
We propose a novel and efficient training scheme for adaptive SimulMT.
Our method outperforms all strong baselines in terms of translation quality and latency.
arXiv Detail & Related papers (2022-04-27T02:40:21Z) - Anticipation-free Training for Simultaneous Translation [70.85761141178597]
Simultaneous translation (SimulMT) speeds up the translation process by starting to translate before the source sentence is completely available.
Existing methods increase latency or introduce adaptive read-write policies for SimulMT models to handle local reordering and improve translation quality.
We propose a new framework that decomposes the translation process into the monotonic translation step and the reordering step.
arXiv Detail & Related papers (2022-01-30T16:29:37Z) - Future-Guided Incremental Transformer for Simultaneous Translation [6.8452940299620435]
Simultaneous translation (ST) starts synchronously while reading source sentences, and is used in many online scenarios.
wait-k policy faces two weaknesses: low training speed caused by the recalculation of hidden states and lack of future source information to guide training.
We propose an incremental Transformer with an average embedding layer (AEL) to accelerate the speed of calculation of hidden states.
arXiv Detail & Related papers (2020-12-23T03:04:49Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z) - Learning Coupled Policies for Simultaneous Machine Translation using
Imitation Learning [85.70547744787]
We present an approach to efficiently learn a simultaneous translation model with coupled programmer-interpreter policies.
Experiments on six language-pairs show our method outperforms strong baselines in terms of translation quality.
arXiv Detail & Related papers (2020-02-11T10:56:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.