TransLaw: Benchmarking Large Language Models in Multi-Agent Simulation of the Collaborative Translation
- URL: http://arxiv.org/abs/2507.00875v1
- Date: Tue, 01 Jul 2025 15:39:26 GMT
- Title: TransLaw: Benchmarking Large Language Models in Multi-Agent Simulation of the Collaborative Translation
- Authors: Xi Xuan, King-kui Sin, Yufei Zhou, Chunyu Kit,
- Abstract summary: TransLaw is a novel multi-agent framework implemented for real-world Hong Kong case law translation.<n>It employs three specialized agents, namely, Translator, Annotator, and Proofreader, to collaboratively produce translations for high accuracy in legal meaning.
- Score: 1.112686067941444
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-agent systems empowered by large language models (LLMs) have demonstrated remarkable capabilities in a wide range of downstream applications, including machine translation. However, the potential of LLMs in translating Hong Kong legal judgments remains uncertain due to challenges such as intricate legal terminology, culturally embedded nuances, and strict linguistic structures. In this work, we introduce TransLaw, a novel multi-agent framework implemented for real-world Hong Kong case law translation. It employs three specialized agents, namely, Translator, Annotator, and Proofreader, to collaboratively produce translations for high accuracy in legal meaning, appropriateness in style, and adequate coherence and cohesion in structure. This framework supports customizable LLM configurations and achieves tremendous cost reduction compared to professional human translation services. We evaluated its performance using 13 open-source and commercial LLMs as agents and obtained interesting findings, including that it surpasses GPT-4o in legal semantic accuracy, structural coherence, and stylistic fidelity, yet trails human experts in contextualizing complex terminology and stylistic naturalness. Our platform website is available at CityUHK, and our bilingual judgment corpus used for the evaluation is available at Hugging Face.
Related papers
- TACTIC: Translation Agents with Cognitive-Theoretic Interactive Collaboration [19.58067098896903]
We propose a cognitively informed multi-agent framework called TACTIC.<n>It comprises six functionally distinct agents that mirror key cognitive processes observed in human translation behavior.<n>Our method consistently achieves state-of-the-art performance.
arXiv Detail & Related papers (2025-06-10T03:22:30Z) - Collaboration is all you need: LLM Assisted Safe Code Translation [4.3764649156831235]
UniTranslator is a framework that re-imagines code translation as a collaborative endeavor among multiple, compact LLMs.<n>By orchestrating the interaction of specialized agents, UniTranslator achieves a level of accuracy and efficiency that rivals larger, monolithic models.
arXiv Detail & Related papers (2025-03-14T09:42:07Z) - SwiLTra-Bench: The Swiss Legal Translation Benchmark [10.2713063405843]
SwiLTra-Bench is a comprehensive benchmark of over 180K aligned Swiss legal translation pairs.<n>Our systematic evaluation reveals that frontier models achieve superior translation performance across all document types.<n>We present SwiLTra-Judge, a specialized LLM evaluation system that aligns best with human expert assessments.
arXiv Detail & Related papers (2025-03-03T10:10:30Z) - Solving the Unsolvable: Translating Case Law in Hong Kong [0.5605104491423386]
The paper critiques the governments and judiciarys sporadic and uncoordinated efforts to translate case law.<n>A proposed solution involves leveraging machine translation technology through a human-machine interactive translation platform.
arXiv Detail & Related papers (2025-01-16T10:17:58Z) - CRAT: A Multi-Agent Framework for Causality-Enhanced Reflective and Retrieval-Augmented Translation with Large Language Models [59.8529196670565]
CRAT is a novel multi-agent translation framework that leverages RAG and causality-enhanced self-reflection to address translation challenges.
Our results show that CRAT significantly improves translation accuracy, particularly in handling context-sensitive terms and emerging vocabulary.
arXiv Detail & Related papers (2024-10-28T14:29:11Z) - LLM-based Translation Inference with Iterative Bilingual Understanding [52.46978502902928]
We propose a novel Iterative Bilingual Understanding Translation method based on the cross-lingual capabilities of large language models (LLMs)<n>The cross-lingual capability of LLMs enables the generation of contextual understanding for both the source and target languages separately.<n>The proposed IBUT outperforms several strong comparison methods.
arXiv Detail & Related papers (2024-10-16T13:21:46Z) - (Perhaps) Beyond Human Translation: Harnessing Multi-Agent Collaboration for Translating Ultra-Long Literary Texts [56.7988577327046]
We introduce TransAgents, a novel multi-agent framework that simulates the roles and collaborative practices of a human translation company.<n>Our findings highlight the potential of multi-agent collaboration in enhancing translation quality, particularly for longer texts.
arXiv Detail & Related papers (2024-05-20T05:55:08Z) - Eliciting the Translation Ability of Large Language Models via Multilingual Finetuning with Translation Instructions [68.01449013641532]
Large-scale Pretrained Language Models (LLMs) have shown strong abilities in multilingual translations.
We present a detailed analysis by finetuning a multilingual pretrained language model, XGLM-7B, to perform multilingual translation.
arXiv Detail & Related papers (2023-05-24T12:00:24Z) - Exploring Human-Like Translation Strategy with Large Language Models [93.49333173279508]
Large language models (LLMs) have demonstrated impressive capabilities in general scenarios.
This work proposes the MAPS framework, which stands for Multi-Aspect Prompting and Selection.
We employ a selection mechanism based on quality estimation to filter out noisy and unhelpful knowledge.
arXiv Detail & Related papers (2023-05-06T19:03:12Z) - Evaluating Multilingual Text Encoders for Unsupervised Cross-Lingual
Retrieval [51.60862829942932]
We present a systematic empirical study focused on the suitability of the state-of-the-art multilingual encoders for cross-lingual document and sentence retrieval tasks.
For sentence-level CLIR, we demonstrate that state-of-the-art performance can be achieved.
However, the peak performance is not met using the general-purpose multilingual text encoders off-the-shelf', but rather relying on their variants that have been further specialized for sentence understanding tasks.
arXiv Detail & Related papers (2021-01-21T00:15:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.