An Investigation of Warning Erroneous Chat Translations in Cross-lingual Communication
- URL: http://arxiv.org/abs/2408.15543v2
- Date: Tue, 5 Nov 2024 04:13:55 GMT
- Title: An Investigation of Warning Erroneous Chat Translations in Cross-lingual Communication
- Authors: Yunmeng Li, Jun Suzuki, Makoto Morishita, Kaori Abe, Kentaro Inui,
- Abstract summary: Machine translation models are still inappropriate for translating chats, despite the popularity of translation software and plug-in applications.
Instead of pursuing a flawless translation system, a more practical approach would be to issue warning messages about potential mistranslations to reduce confusion.
This paper tackles to investigate this question and demonstrates the warning messages' contribution to making chat translation systems effective.
- Score: 35.69695355173317
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine translation models are still inappropriate for translating chats, despite the popularity of translation software and plug-in applications. The complexity of dialogues poses significant challenges and can hinder crosslingual communication. Instead of pursuing a flawless translation system, a more practical approach would be to issue warning messages about potential mistranslations to reduce confusion. However, it is still unclear how individuals perceive these warning messages and whether they benefit the crowd. This paper tackles to investigate this question and demonstrates the warning messages' contribution to making chat translation systems effective.
Related papers
- A Context-aware Framework for Translation-mediated Conversations [29.169155271343083]
We present a framework to improve large language model-based translation systems by incorporating contextual information in bilingual conversational settings.
We validate both components of our framework on two task-oriented domains: customer chat and user-assistant interaction.
Our framework consistently results in better translations than state-of-the-art systems like GPT-4o and TowerInstruct.
arXiv Detail & Related papers (2024-12-05T14:41:05Z) - Do Prompts Really Prompt? Exploring the Prompt Understanding Capability of Whisper [51.12146889808824]
This research explores how the information of prompts interacts with the high-performing speech recognition model, Whisper.
Our results unexpectedly show that Whisper may not understand the textual prompts in a human-expected way.
It is also noted that English prompts generally outperform Mandarin ones on datasets of both languages.
arXiv Detail & Related papers (2024-06-09T14:44:59Z) - Crossing the Threshold: Idiomatic Machine Translation through Retrieval
Augmentation and Loss Weighting [66.02718577386426]
We provide a simple characterization of idiomatic translation and related issues.
We conduct a synthetic experiment revealing a tipping point at which transformer-based machine translation models correctly default to idiomatic translations.
To improve translation of natural idioms, we introduce two straightforward yet effective techniques.
arXiv Detail & Related papers (2023-10-10T23:47:25Z) - Mitigating Hallucinations and Off-target Machine Translation with
Source-Contrastive and Language-Contrastive Decoding [53.84948040596055]
We introduce two related methods to mitigate failure cases with a modified decoding objective.
Experiments on the massively multilingual models M2M-100 (418M) and SMaLL-100 show that these methods suppress hallucinations and off-target translations.
arXiv Detail & Related papers (2023-09-13T17:15:27Z) - Chat Translation Error Detection for Assisting Cross-lingual
Communications [35.09508360315392]
We train an error detector as the baseline of the system and construct a new Japanese-English bilingual chat corpus, BPersona-chat.
The error detector can serve as an encouraging foundation for more advanced erroneous translation detection systems.
arXiv Detail & Related papers (2023-08-02T09:38:29Z) - Lost In Translation: Generating Adversarial Examples Robust to
Round-Trip Translation [66.33340583035374]
We present a comprehensive study on the robustness of current text adversarial attacks to round-trip translation.
We demonstrate that 6 state-of-the-art text-based adversarial attacks do not maintain their efficacy after round-trip translation.
We introduce an intervention-based solution to this problem, by integrating Machine Translation into the process of adversarial example generation.
arXiv Detail & Related papers (2023-07-24T04:29:43Z) - Translate your gibberish: black-box adversarial attack on machine
translation systems [0.0]
We present a simple approach to fool state-of-the-art machine translation tools in the task of translation from Russian to English and vice versa.
We show that many online translation tools, such as Google, DeepL, and Yandex, may both produce wrong or offensive translations for nonsensical adversarial input queries.
This vulnerability may interfere with understanding a new language and simply worsen the user's experience while using machine translation systems.
arXiv Detail & Related papers (2023-03-20T09:52:52Z) - Interactive-Chain-Prompting: Ambiguity Resolution for Crosslingual
Conditional Generation with Interaction [38.73550742775257]
A source query in one language may yield several translation options in another language without any extra context.
We propose a novel method interactive-chain prompting that reduces translations into a list of subproblems addressing ambiguities.
We create a dataset exhibiting different linguistic phenomena which leads to ambiguities at inference for four languages.
arXiv Detail & Related papers (2023-01-24T21:08:13Z) - A Semi-supervised Approach for a Better Translation of Sentiment in
Dialectical Arabic UGT [2.6763498831034034]
We introduce a semi-supervised approach that exploits both monolingual and parallel data for training an NMT system.
We will show that our proposed system can significantly help with correcting sentiment errors detected in the online translation of dialectical Arabic UGT.
arXiv Detail & Related papers (2022-10-21T11:55:55Z) - Self-Supervised Representations Improve End-to-End Speech Translation [57.641761472372814]
We show that self-supervised pre-trained features can consistently improve the translation performance.
Cross-lingual transfer allows to extend to a variety of languages without or with little tuning.
arXiv Detail & Related papers (2020-06-22T10:28:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.