IMS' Systems for the IWSLT 2021 Low-Resource Speech Translation Task
- URL: http://arxiv.org/abs/2106.16055v1
- Date: Wed, 30 Jun 2021 13:29:19 GMT
- Title: IMS' Systems for the IWSLT 2021 Low-Resource Speech Translation Task
- Authors: Pavel Denisov, Manuel Mager, Ngoc Thang Vu
- Abstract summary: This paper describes the submission to the IWSLT 2021 Low-Resource Speech Translation Shared Task by IMS team.
We utilize state-of-the-art models combined with several data augmentation, multi-task and transfer learning approaches for the automatic speech recognition (ASR) and machine translation (MT) steps of our cascaded system.
- Score: 38.899667657333595
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: This paper describes the submission to the IWSLT 2021 Low-Resource Speech
Translation Shared Task by IMS team. We utilize state-of-the-art models
combined with several data augmentation, multi-task and transfer learning
approaches for the automatic speech recognition (ASR) and machine translation
(MT) steps of our cascaded system. Moreover, we also explore the feasibility of
a full end-to-end speech translation (ST) model in the case of very constrained
amount of ground truth labeled data. Our best system achieves the best
performance among all submitted systems for Congolese Swahili to English and
French with BLEU scores 7.7 and 13.7 respectively, and the second best result
for Coastal Swahili to English with BLEU score 14.9.
Related papers
- NAVER LABS Europe's Multilingual Speech Translation Systems for the
IWSLT 2023 Low-Resource Track [78.80683163990446]
This paper presents NAVER LABS Europe's systems for Tamasheq-French and Quechua-Spanish speech translation in the IWSLT 2023 Low-Resource track.
Our work attempts to maximize translation quality in low-resource settings using multilingual parameter-efficient solutions.
arXiv Detail & Related papers (2023-06-13T13:22:30Z) - KIT's Multilingual Speech Translation System for IWSLT 2023 [58.5152569458259]
We describe our speech translation system for the multilingual track of IWSLT 2023.
The task requires translation into 10 languages of varying amounts of resources.
Our cascaded speech system substantially outperforms its end-to-end counterpart on scientific talk translation.
arXiv Detail & Related papers (2023-06-08T16:13:20Z) - Tencent AI Lab - Shanghai Jiao Tong University Low-Resource Translation
System for the WMT22 Translation Task [49.916963624249355]
This paper describes Tencent AI Lab - Shanghai Jiao Tong University (TAL-SJTU) Low-Resource Translation systems for the WMT22 shared task.
We participate in the general translation task on English$Leftrightarrow$Livonian.
Our system is based on M2M100 with novel techniques that adapt it to the target language pair.
arXiv Detail & Related papers (2022-10-17T04:34:09Z) - The YiTrans End-to-End Speech Translation System for IWSLT 2022 Offline
Shared Task [92.5087402621697]
This paper describes the submission of our end-to-end YiTrans speech translation system for the IWSLT 2022 offline task.
The YiTrans system is built on large-scale pre-trained encoder-decoder models.
Our final submissions rank first on English-German and English-Chinese end-to-end systems in terms of the automatic evaluation metric.
arXiv Detail & Related papers (2022-06-12T16:13:01Z) - The USYD-JD Speech Translation System for IWSLT 2021 [85.64797317290349]
This paper describes the University of Sydney& JD's joint submission of the IWSLT 2021 low resource speech translation task.
We trained our models with the officially provided ASR and MT datasets.
To achieve better translation performance, we explored the most recent effective strategies, including back translation, knowledge distillation, multi-feature reranking and transductive finetuning.
arXiv Detail & Related papers (2021-07-24T09:53:34Z) - FST: the FAIR Speech Translation System for the IWSLT21 Multilingual
Shared Task [36.51221186190272]
We describe our end-to-end multilingual speech translation system submitted to the IWSLT 2021 evaluation campaign.
Our system is built by leveraging transfer learning across modalities, tasks and languages.
arXiv Detail & Related papers (2021-07-14T19:43:44Z) - The Volctrans Neural Speech Translation System for IWSLT 2021 [26.058205594318405]
This paper describes the systems submitted to IWSLT 2021 by the Volctrans team.
For offline speech translation, our best end-to-end model achieves 8.1 BLEU improvements over the benchmark.
For text-to-text simultaneous translation, we explore the best practice to optimize the wait-k model.
arXiv Detail & Related papers (2021-05-16T00:11:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.