Controlling Formality in Low-Resource NMT with Domain Adaptation and
Re-Ranking: SLT-CDT-UoS at IWSLT2022
- URL: http://arxiv.org/abs/2205.05990v1
- Date: Thu, 12 May 2022 09:54:17 GMT
- Title: Controlling Formality in Low-Resource NMT with Domain Adaptation and
Re-Ranking: SLT-CDT-UoS at IWSLT2022
- Authors: Sebastian T. Vincent, Lo\"ic Barrault, Carolina Scarton
- Abstract summary: This paper describes the SLT-CDT-UoS group's submission to the first Special Task on Formality Control for Spoken Language Translation.
Our efforts were split between two fronts: data engineering and altering the objective function for best hypothesis selection.
On the test sets for English-to-German and English-to-Spanish, we achieved an average accuracy of.935 within the constrained setting and.995 within unconstrained setting.
- Score: 4.348327991071386
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper describes the SLT-CDT-UoS group's submission to the first Special
Task on Formality Control for Spoken Language Translation, part of the IWSLT
2022 Evaluation Campaign. Our efforts were split between two fronts: data
engineering and altering the objective function for best hypothesis selection.
We used language-independent methods to extract formal and informal sentence
pairs from the provided corpora; using English as a pivot language, we
propagated formality annotations to languages treated as zero-shot in the task;
we also further improved formality controlling with a hypothesis re-ranking
approach. On the test sets for English-to-German and English-to-Spanish, we
achieved an average accuracy of .935 within the constrained setting and .995
within unconstrained setting. In a zero-shot setting for English-to-Russian and
English-to-Italian, we scored average accuracy of .590 for constrained setting
and .659 for unconstrained.
Related papers
- The Power of Question Translation Training in Multilingual Reasoning: Broadened Scope and Deepened Insights [108.40766216456413]
We propose a question alignment framework to bridge the gap between large language models' English and non-English performance.
Experiment results show it can boost multilingual performance across diverse reasoning scenarios, model families, and sizes.
We analyze representation space, generated response and data scales, and reveal how question translation training strengthens language alignment within LLMs.
arXiv Detail & Related papers (2024-05-02T14:49:50Z) - Prefix Text as a Yarn: Eliciting Non-English Alignment in Foundation Language Model [50.339632513018934]
supervised fine-tuning (SFT) has been a straightforward approach for tailoring the output of foundation large language model (LLM) to specific preferences.
We critically examine this hypothesis within the scope of cross-lingual generation tasks.
We introduce a novel training-free alignment method named PreTTY, which employs minimal task-related prior tokens.
arXiv Detail & Related papers (2024-04-25T17:19:36Z) - Breaking the Language Barrier: Can Direct Inference Outperform
Pre-Translation in Multilingual LLM Applications? [17.828943682809882]
This study re-evaluates the need for pre-translation in the context of PaLM2 models.
PaLM2-L consistently outperforms pre-translation in 94 out of 108 languages.
arXiv Detail & Related papers (2024-03-04T14:01:11Z) - NAVER LABS Europe's Multilingual Speech Translation Systems for the
IWSLT 2023 Low-Resource Track [78.80683163990446]
This paper presents NAVER LABS Europe's systems for Tamasheq-French and Quechua-Spanish speech translation in the IWSLT 2023 Low-Resource track.
Our work attempts to maximize translation quality in low-resource settings using multilingual parameter-efficient solutions.
arXiv Detail & Related papers (2023-06-13T13:22:30Z) - KIT's Multilingual Speech Translation System for IWSLT 2023 [58.5152569458259]
We describe our speech translation system for the multilingual track of IWSLT 2023.
The task requires translation into 10 languages of varying amounts of resources.
Our cascaded speech system substantially outperforms its end-to-end counterpart on scientific talk translation.
arXiv Detail & Related papers (2023-06-08T16:13:20Z) - Strategies for improving low resource speech to text translation relying
on pre-trained ASR models [59.90106959717875]
This paper presents techniques and findings for improving the performance of low-resource speech to text translation (ST)
We conducted experiments on both simulated and real-low resource setups, on language pairs English - Portuguese, and Tamasheq - French respectively.
arXiv Detail & Related papers (2023-05-31T21:58:07Z) - The Interpreter Understands Your Meaning: End-to-end Spoken Language
Understanding Aided by Speech Translation [13.352795145385645]
Speech translation (ST) is a good means of pretraining speech models for end-to-end spoken language understanding.
We show that our models reach higher performance over baselines on monolingual and multilingual intent classification.
We also create new benchmark datasets for speech summarization and low-resource/zero-shot transfer from English to French or Spanish.
arXiv Detail & Related papers (2023-05-16T17:53:03Z) - Tencent AI Lab - Shanghai Jiao Tong University Low-Resource Translation
System for the WMT22 Translation Task [49.916963624249355]
This paper describes Tencent AI Lab - Shanghai Jiao Tong University (TAL-SJTU) Low-Resource Translation systems for the WMT22 shared task.
We participate in the general translation task on English$Leftrightarrow$Livonian.
Our system is based on M2M100 with novel techniques that adapt it to the target language pair.
arXiv Detail & Related papers (2022-10-17T04:34:09Z) - Controlling Translation Formality Using Pre-trained Multilingual
Language Models [19.465727478912072]
This paper describes the University of Maryland's submission to the Special Task on Formality Control for Spoken Language Translation at iwslt.
We investigate to what extent this problem can be addressed with a textitsingle multilingual model.
Results show that this strategy can approach the translation quality and formality control achieved by dedicated translation models.
arXiv Detail & Related papers (2022-05-13T13:47:28Z) - Zero-Shot Cross-Lingual Transfer in Legal Domain Using Transformer
models [0.0]
We study zero-shot cross-lingual transfer from English to French and German under Multi-Label Text Classification.
We extend EURLEX57K dataset, the English dataset for topic classification of legal documents, with French and German official translation.
We find that Language model finetuning of multi-lingual pre-trained model (M-DistilBERT, M-BERT) leads to 32.0-34.94%, 76.15-87.54% relative improvement on French and German test sets.
arXiv Detail & Related papers (2021-11-28T16:25:04Z) - Ensemble Fine-tuned mBERT for Translation Quality Estimation [0.0]
In this paper, we discuss our submission to the WMT 2021 QE Shared Task.
Our proposed system is an ensemble of multilingual BERT (mBERT)-based regression models.
It demonstrates comparable performance with respect to the Pearson's correlation and beats the baseline system in MAE/ RMSE for several language pairs.
arXiv Detail & Related papers (2021-09-08T20:13:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.