SlangDIT: Benchmarking LLMs in Interpretative Slang Translation
- URL: http://arxiv.org/abs/2505.14181v1
- Date: Tue, 20 May 2025 10:37:34 GMT
- Title: SlangDIT: Benchmarking LLMs in Interpretative Slang Translation
- Authors: Yunlong Liang, Fandong Meng, Jiaan Wang, Jie Zhou,
- Abstract summary: This paper introduces the interpretative slang translation task (named SlangDIT)<n>It consists of three sub-tasks: slang detection, cross-lingual slang explanation, and slang translation within the current context.<n>Based on the benchmark, we propose a deep thinking model, named SlangOWL. It firstly identifies whether the sentence contains a slang, and then judges whether the slang is polysemous and analyze its possible meaning.
- Score: 89.48208612476068
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The challenge of slang translation lies in capturing context-dependent semantic extensions, as slang terms often convey meanings beyond their literal interpretation. While slang detection, explanation, and translation have been studied as isolated tasks in the era of large language models (LLMs), their intrinsic interdependence remains underexplored. The main reason is lacking of a benchmark where the two tasks can be a prerequisite for the third one, which can facilitate idiomatic translation. In this paper, we introduce the interpretative slang translation task (named SlangDIT) consisting of three sub-tasks: slang detection, cross-lingual slang explanation, and slang translation within the current context, aiming to generate more accurate translation with the help of slang detection and slang explanation. To this end, we construct a SlangDIT dataset, containing over 25k English-Chinese sentence pairs. Each source sentence mentions at least one slang term and is labeled with corresponding cross-lingual slang explanation. Based on the benchmark, we propose a deep thinking model, named SlangOWL. It firstly identifies whether the sentence contains a slang, and then judges whether the slang is polysemous and analyze its possible meaning. Further, the SlangOWL provides the best explanation of the slang term targeting on the current context. Finally, according to the whole thought, the SlangOWL offers a suitable translation. Our experiments on LLMs (\emph{e.g.}, Qwen2.5 and LLama-3.1), show that our deep thinking approach indeed enhances the performance of LLMs where the proposed SLangOWL significantly surpasses the vanilla models and supervised fine-tuned models without thinking.
Related papers
- Please Translate Again: Two Simple Experiments on Whether Human-Like Reasoning Helps Translation [7.376832526909754]
Large Language Models (LLMs) demonstrate strong reasoning capabilities for many tasks, often by explicitly decomposing the task via Chain-of-Thought (CoT) reasoning.<n>textitTranslating Step-by-stepcitepbriakou2024translating, for instance, introduces a multi-step prompt with decomposition and refinement of translation with LLMs.<n>We show that simply prompting LLMs to translate again'' yields even better results than human-like step-by-step prompting.
arXiv Detail & Related papers (2025-06-05T00:04:39Z) - Deep Reasoning Translation via Reinforcement Learning [77.41383117199227]
We introduce DeepTrans, a deep reasoning translation model that learns free translation via reinforcement learning.<n>Given the source sentences, the reward model teaches the deep translation model how to think and free-translate them during reinforcement learning.<n> Experimental results show that DeepTrans improves performance by 16.3% in literature translation.
arXiv Detail & Related papers (2025-04-14T12:40:39Z) - DRT: Deep Reasoning Translation via Long Chain-of-Thought [89.48208612476068]
In this paper, we introduce DRT, an attempt to bring the success of long CoT to neural machine translation (MT)<n>We first mine sentences containing similes or metaphors from existing literature books, and then develop a multi-agent framework to translate these sentences via long thought.<n>Using Qwen2.5 and LLama-3.1 as the backbones, DRT models can learn the thought process during machine translation.
arXiv Detail & Related papers (2024-12-23T11:55:33Z) - Improving LLM Abilities in Idiomatic Translation [2.8692611791027893]
For language models (LLMs) like NLLB and GPT, translating idioms remains a challenge.<n>Our goal is to enhance translation fidelity by improving LLM processing of idiomatic language.<n>This has a significant social impact, as it preserves cultural nuances and ensures translated texts retain intent and emotional resonance.
arXiv Detail & Related papers (2024-07-03T21:34:26Z) - Incremental Comprehension of Garden-Path Sentences by Large Language Models: Semantic Interpretation, Syntactic Re-Analysis, and Attention [11.073959609358088]
We investigate the processing of garden-path sentences and the fate of lingering misinterpretations using four large language models.
The overall goal is to evaluate whether humans and LLMs are aligned in their processing of garden-path sentences.
Experiments show promising alignment between humans and LLMs in the processing of garden-path sentences.
arXiv Detail & Related papers (2024-05-25T03:36:13Z) - That was the last straw, we need more: Are Translation Systems Sensitive
to Disambiguating Context? [64.38544995251642]
We study semantic ambiguities that exist in the source (English in this work) itself.
We focus on idioms that are open to both literal and figurative interpretations.
We find that current MT models consistently translate English idioms literally, even when the context suggests a figurative interpretation.
arXiv Detail & Related papers (2023-10-23T06:38:49Z) - Towards Effective Disambiguation for Machine Translation with Large
Language Models [65.80775710657672]
We study the capabilities of large language models to translate "ambiguous sentences"
Experiments show that our methods can match or outperform state-of-the-art systems such as DeepL and NLLB in four out of five language directions.
arXiv Detail & Related papers (2023-09-20T22:22:52Z) - Translate to Disambiguate: Zero-shot Multilingual Word Sense
Disambiguation with Pretrained Language Models [67.19567060894563]
Pretrained Language Models (PLMs) learn rich cross-lingual knowledge and can be finetuned to perform well on diverse tasks.
We present a new study investigating how well PLMs capture cross-lingual word sense with Contextual Word-Level Translation (C-WLT)
We find that as the model size increases, PLMs encode more cross-lingual word sense knowledge and better use context to improve WLT performance.
arXiv Detail & Related papers (2023-04-26T19:55:52Z) - Semantically Informed Slang Interpretation [2.9097456604613745]
We propose a semantically informed slang interpretation (SSI) framework that considers jointly the contextual and semantic appropriateness of a candidate interpretation for a query slang.
We show how the same framework can be applied to enhancing machine translation of slang from English to other languages.
arXiv Detail & Related papers (2022-05-02T01:51:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.