Tilde at WMT 2020: News Task Systems
- URL: http://arxiv.org/abs/2010.15423v1
- Date: Thu, 29 Oct 2020 08:59:37 GMT
- Title: Tilde at WMT 2020: News Task Systems
- Authors: Rihards Kri\v{s}lauks, M\=arcis Pinnis
- Abstract summary: This paper describes Tilde's submission to the WMT 2020 shared task on news translation for both directions of the English-Polish language pair.
We build our baseline systems to be morphologically motivated sub-word unit-based Transformer base models.
Our final models are ensembles of Transformer base and Transformer big models that feature right-to-left re-ranking.
- Score: 0.38073142980733
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper describes Tilde's submission to the WMT2020 shared task on news
translation for both directions of the English-Polish language pair in both the
constrained and the unconstrained tracks. We follow our submissions from the
previous years and build our baseline systems to be morphologically motivated
sub-word unit-based Transformer base models that we train using the Marian
machine translation toolkit. Additionally, we experiment with different
parallel and monolingual data selection schemes, as well as sampled
back-translation. Our final models are ensembles of Transformer base and
Transformer big models that feature right-to-left re-ranking.
Related papers
- HW-TSC's Submission to the CCMT 2024 Machine Translation Tasks [12.841065384808733]
We participate in the bilingual machine translation task and multi-domain machine translation task.
For these two translation tasks, we use training strategies such as regularized dropout, bidirectional training, data diversification, forward translation, back translation, alternated training, curriculum learning, and transductive ensemble learning.
arXiv Detail & Related papers (2024-09-23T09:20:19Z) - Choose the Final Translation from NMT and LLM hypotheses Using MBR Decoding: HW-TSC's Submission to the WMT24 General MT Shared Task [9.819139035652137]
This paper presents the submission of Huawei Translate Services Center (HW-TSC) to the WMT24 general machine translation (MT) shared task.
We use training strategies such as regularized dropout, bidirectional training, data diversification, forward translation, back translation, alternated training, curriculum learning, and transductive ensemble learning to train the neural machine translation (NMT) model.
arXiv Detail & Related papers (2024-09-23T08:25:37Z) - Summer: WeChat Neural Machine Translation Systems for the WMT22
Biomedical Translation Task [54.63368889359441]
This paper introduces WeChat's participation in WMT 2022 shared biomedical translation task on Chinese to English.
Our systems are based on the Transformer, and use several different Transformer structures to improve the quality of translation.
Our Chinese$to$English system, named Summer, achieves the highest BLEU score among all submissions.
arXiv Detail & Related papers (2022-11-28T03:10:50Z) - Improved Data Augmentation for Translation Suggestion [28.672227843541656]
This paper introduces the system used in our submission to the WMT'22 Translation Suggestion shared task.
We use three strategies to construct synthetic data from parallel corpora to compensate for the lack of supervised data.
We rank second and third on the English-German and English-Chinese bidirectional tasks, respectively.
arXiv Detail & Related papers (2022-10-12T12:46:43Z) - The YiTrans End-to-End Speech Translation System for IWSLT 2022 Offline
Shared Task [92.5087402621697]
This paper describes the submission of our end-to-end YiTrans speech translation system for the IWSLT 2022 offline task.
The YiTrans system is built on large-scale pre-trained encoder-decoder models.
Our final submissions rank first on English-German and English-Chinese end-to-end systems in terms of the automatic evaluation metric.
arXiv Detail & Related papers (2022-06-12T16:13:01Z) - XLM-T: Scaling up Multilingual Machine Translation with Pretrained
Cross-lingual Transformer Encoders [89.0059978016914]
We present XLM-T, which initializes the model with an off-the-shelf pretrained cross-lingual Transformer and fine-tunes it with multilingual parallel data.
This simple method achieves significant improvements on a WMT dataset with 10 language pairs and the OPUS-100 corpus with 94 pairs.
arXiv Detail & Related papers (2020-12-31T11:16:51Z) - DiDi's Machine Translation System for WMT2020 [51.296629834996246]
We participate in the translation direction of Chinese->English.
In this direction, we use the Transformer as our baseline model.
As a result, our submission achieves a BLEU score of $36.6$ in Chinese->English.
arXiv Detail & Related papers (2020-10-16T06:25:48Z) - SJTU-NICT's Supervised and Unsupervised Neural Machine Translation
Systems for the WMT20 News Translation Task [111.91077204077817]
We participated in four translation directions of three language pairs: English-Chinese, English-Polish, and German-Upper Sorbian.
Based on different conditions of language pairs, we have experimented with diverse neural machine translation (NMT) techniques.
In our submissions, the primary systems won the first place on English to Chinese, Polish to English, and German to Upper Sorbian translation directions.
arXiv Detail & Related papers (2020-10-11T00:40:05Z) - Learning Contextualized Sentence Representations for Document-Level
Neural Machine Translation [59.191079800436114]
Document-level machine translation incorporates inter-sentential dependencies into the translation of a source sentence.
We propose a new framework to model cross-sentence dependencies by training neural machine translation (NMT) to predict both the target translation and surrounding sentences of a source sentence.
arXiv Detail & Related papers (2020-03-30T03:38:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.