WeChat Neural Machine Translation Systems for WMT21
- URL: http://arxiv.org/abs/2108.02401v1
- Date: Thu, 5 Aug 2021 06:38:48 GMT
- Title: WeChat Neural Machine Translation Systems for WMT21
- Authors: Xianfeng Zeng, Yijin Liu, Ernan Li, Qiu Ran, Fandong Meng, Peng Li,
Jinan Xu and Jie Zhou
- Abstract summary: This paper introduces AI's participation in WMT 2021 WeChat shared news translation task on English->Chinese, English->Japanese, Japanese->English and English->German.
We employ data filtering, large-scale synthetic data generation, advanced finetuning approaches, and boosted Self-BLEU based model ensemble.
Our constrained systems achieve 36.9, 46.9, 27.8 and 31.3 case-sensitive BLEU scores on English->Chinese, English->Japanese, Japanese->English and English->German, respectively.
- Score: 22.51171167457826
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces WeChat AI's participation in WMT 2021 shared news
translation task on English->Chinese, English->Japanese, Japanese->English and
English->German. Our systems are based on the Transformer (Vaswani et al.,
2017) with several novel and effective variants. In our experiments, we employ
data filtering, large-scale synthetic data generation (i.e., back-translation,
knowledge distillation, forward-translation, iterative in-domain knowledge
transfer), advanced finetuning approaches, and boosted Self-BLEU based model
ensemble. Our constrained systems achieve 36.9, 46.9, 27.8 and 31.3
case-sensitive BLEU scores on English->Chinese, English->Japanese,
Japanese->English and English->German, respectively. The BLEU scores of
English->Chinese, English->Japanese and Japanese->English are the highest among
all submissions, and that of English->German is the highest among all
constrained submissions.
Related papers
- Summer: WeChat Neural Machine Translation Systems for the WMT22
Biomedical Translation Task [54.63368889359441]
This paper introduces WeChat's participation in WMT 2022 shared biomedical translation task on Chinese to English.
Our systems are based on the Transformer, and use several different Transformer structures to improve the quality of translation.
Our Chinese$to$English system, named Summer, achieves the highest BLEU score among all submissions.
arXiv Detail & Related papers (2022-11-28T03:10:50Z) - BJTU-WeChat's Systems for the WMT22 Chat Translation Task [66.81525961469494]
This paper introduces the joint submission of the Beijing Jiaotong University and WeChat AI to the WMT'22 chat translation task for English-German.
Based on the Transformer, we apply several effective variants.
Our systems achieve 0.810 and 0.946 COMET scores.
arXiv Detail & Related papers (2022-11-28T02:35:04Z) - Tencent AI Lab - Shanghai Jiao Tong University Low-Resource Translation
System for the WMT22 Translation Task [49.916963624249355]
This paper describes Tencent AI Lab - Shanghai Jiao Tong University (TAL-SJTU) Low-Resource Translation systems for the WMT22 shared task.
We participate in the general translation task on English$Leftrightarrow$Livonian.
Our system is based on M2M100 with novel techniques that adapt it to the target language pair.
arXiv Detail & Related papers (2022-10-17T04:34:09Z) - The YiTrans End-to-End Speech Translation System for IWSLT 2022 Offline
Shared Task [92.5087402621697]
This paper describes the submission of our end-to-end YiTrans speech translation system for the IWSLT 2022 offline task.
The YiTrans system is built on large-scale pre-trained encoder-decoder models.
Our final submissions rank first on English-German and English-Chinese end-to-end systems in terms of the automatic evaluation metric.
arXiv Detail & Related papers (2022-06-12T16:13:01Z) - DiDi's Machine Translation System for WMT2020 [51.296629834996246]
We participate in the translation direction of Chinese->English.
In this direction, we use the Transformer as our baseline model.
As a result, our submission achieves a BLEU score of $36.6$ in Chinese->English.
arXiv Detail & Related papers (2020-10-16T06:25:48Z) - WeChat Neural Machine Translation Systems for WMT20 [61.03013964996131]
Our system is based on the Transformer with effective variants and the DTMT architecture.
In our experiments, we employ data selection, several synthetic data generation approaches, advanced finetuning approaches and self-bleu based model ensemble.
Our constrained Chinese to English system achieves 36.9 case-sensitive BLEU score, which is the highest among all submissions.
arXiv Detail & Related papers (2020-10-01T08:15:09Z) - Neural Machine Translation for Low-Resourced Indian Languages [4.726777092009554]
Machine translation is an effective approach to convert text to a different language without any human involvement.
In this paper, we have applied NMT on two of the most morphological rich Indian languages, i.e. English-Tamil and English-Malayalam.
We proposed a novel NMT model using Multihead self-attention along with pre-trained Byte-Pair-Encoded (BPE) and MultiBPE embeddings to develop an efficient translation system.
arXiv Detail & Related papers (2020-04-19T17:29:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.