Time-Aware Ancient Chinese Text Translation and Inference
- URL: http://arxiv.org/abs/2107.03179v1
- Date: Wed, 7 Jul 2021 12:23:52 GMT
- Title: Time-Aware Ancient Chinese Text Translation and Inference
- Authors: Ernie Chang, Yow-Ting Shiue, Hui-Syuan Yeh, Vera Demberg
- Abstract summary: We aim to address the challenges surrounding the translation of ancient Chinese text.
The linguistic gap due to the difference in eras results in translations that are poor in quality.
Most translations are missing the contextual information that is often very crucial to understanding the text.
- Score: 6.787414471399024
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we aim to address the challenges surrounding the translation
of ancient Chinese text: (1) The linguistic gap due to the difference in eras
results in translations that are poor in quality, and (2) most translations are
missing the contextual information that is often very crucial to understanding
the text. To this end, we improve upon past translation techniques by proposing
the following: We reframe the task as a multi-label prediction task where the
model predicts both the translation and its particular era. We observe that
this helps to bridge the linguistic gap as chronological context is also used
as auxiliary information. % As a natural step of generalization, we pivot on
the modern Chinese translations to generate multilingual outputs. %We show
experimentally the efficacy of our framework in producing quality translation
outputs and also validate our framework on a collected task-specific parallel
corpus. We validate our framework on a parallel corpus annotated with
chronology information and show experimentally its efficacy in producing
quality translation outputs. We release both the code and the data
https://github.com/orina1123/time-aware-ancient-text-translation for future
research.
Related papers
- Mitigating Translationese in Low-resource Languages: The Storyboard Approach [9.676710061071809]
We propose a novel approach for data collection by leveraging storyboards to elicit more fluent and natural sentences.
Our method involves presenting native speakers with visual stimuli in the form of storyboards and collecting their descriptions without direct exposure to the source text.
We conducted a comprehensive evaluation comparing our storyboard-based approach with traditional text translation-based methods in terms of accuracy and fluency.
arXiv Detail & Related papers (2024-07-14T10:47:03Z) - Challenges in Context-Aware Neural Machine Translation [39.89082986080746]
Context-aware neural machine translation involves leveraging information beyond sentence-level context to resolve discourse dependencies.
Despite well-reasoned intuitions, most context-aware translation models show only modest improvements over sentence-level systems.
We propose a more realistic setting for document-level translation, called paragraph-to-paragraph (para2para) translation.
arXiv Detail & Related papers (2023-05-23T07:08:18Z) - HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - Improving End-to-End Text Image Translation From the Auxiliary Text
Translation Task [26.046624228278528]
We propose a novel text translation enhanced text image translation, which trains the end-to-end model with text translation as an auxiliary task.
By sharing model parameters and multi-task training, our model is able to take full advantage of easily-available large-scale text parallel corpus.
arXiv Detail & Related papers (2022-10-08T02:35:45Z) - Towards Debiasing Translation Artifacts [15.991970288297443]
We propose a novel approach to reducing translationese by extending an established bias-removal technique.
We use the Iterative Null-space Projection (INLP) algorithm, and show by measuring classification accuracy before and after debiasing, that translationese is reduced at both sentence and word level.
To the best of our knowledge, this is the first study to debias translationese as represented in latent embedding space.
arXiv Detail & Related papers (2022-05-16T21:46:51Z) - ChrEnTranslate: Cherokee-English Machine Translation Demo with Quality
Estimation and Corrective Feedback [70.5469946314539]
ChrEnTranslate is an online machine translation demonstration system for translation between English and an endangered language Cherokee.
It supports both statistical and neural translation models as well as provides quality estimation to inform users of reliability.
arXiv Detail & Related papers (2021-07-30T17:58:54Z) - Improving Speech Translation by Understanding and Learning from the
Auxiliary Text Translation Task [26.703809355057224]
We conduct a detailed analysis to understand the impact of the auxiliary task on the primary task within the multitask learning framework.
Our analysis confirms that multitask learning tends to generate similar decoder representations from different modalities.
Inspired by these findings, we propose three methods to improve translation quality.
arXiv Detail & Related papers (2021-07-12T23:53:40Z) - On the Language Coverage Bias for Neural Machine Translation [81.81456880770762]
Language coverage bias is important for neural machine translation (NMT) because the target-original training data is not well exploited in current practice.
By carefully designing experiments, we provide comprehensive analyses of the language coverage bias in the training data.
We propose two simple and effective approaches to alleviate the language coverage bias problem.
arXiv Detail & Related papers (2021-06-07T01:55:34Z) - Translation Artifacts in Cross-lingual Transfer Learning [51.66536640084888]
We show that machine translation can introduce subtle artifacts that have a notable impact in existing cross-lingual models.
In natural language inference, translating the premise and the hypothesis independently can reduce the lexical overlap between them.
We also improve the state-of-the-art in XNLI for the translate-test and zero-shot approaches by 4.3 and 2.8 points, respectively.
arXiv Detail & Related papers (2020-04-09T17:54:30Z) - Sign Language Transformers: Joint End-to-end Sign Language Recognition
and Translation [59.38247587308604]
We introduce a novel transformer based architecture that jointly learns Continuous Sign Language Recognition and Translation.
We evaluate the recognition and translation performances of our approaches on the challenging RWTH-PHOENIX-Weather-2014T dataset.
Our translation networks outperform both sign video to spoken language and gloss to spoken language translation models.
arXiv Detail & Related papers (2020-03-30T21:35:09Z) - Learning Coupled Policies for Simultaneous Machine Translation using
Imitation Learning [85.70547744787]
We present an approach to efficiently learn a simultaneous translation model with coupled programmer-interpreter policies.
Experiments on six language-pairs show our method outperforms strong baselines in terms of translation quality.
arXiv Detail & Related papers (2020-02-11T10:56:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.