Sentence Level Human Translation Quality Estimation with Attention-based
Neural Networks
- URL: http://arxiv.org/abs/2003.06381v1
- Date: Fri, 13 Mar 2020 16:57:55 GMT
- Title: Sentence Level Human Translation Quality Estimation with Attention-based
Neural Networks
- Authors: Yu Yuan, Serge Sharoff
- Abstract summary: This paper explores the use of Deep Learning methods for automatic estimation of quality of human translations.
Empirical results on a large human annotated dataset show that the neural model outperforms feature-based methods significantly.
- Score: 0.30458514384586394
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper explores the use of Deep Learning methods for automatic estimation
of quality of human translations. Automatic estimation can provide useful
feedback for translation teaching, examination and quality control.
Conventional methods for solving this task rely on manually engineered features
and external knowledge. This paper presents an end-to-end neural model without
feature engineering, incorporating a cross attention mechanism to detect which
parts in sentence pairs are most relevant for assessing quality. Another
contribution concerns of prediction of fine-grained scores for measuring
different aspects of translation quality. Empirical results on a large human
annotated dataset show that the neural model outperforms feature-based methods
significantly. The dataset and the tools are available.
Related papers
- An Empirical Study on the Robustness of Massively Multilingual Neural Machine Translation [40.08063412966712]
Massively multilingual neural machine translation (MMNMT) has been proven to enhance the translation quality of low-resource languages.
We create a robustness evaluation benchmark dataset for Indonesian-Chinese translation.
This dataset is automatically translated into Chinese using four NLLB-200 models of different sizes.
arXiv Detail & Related papers (2024-05-13T12:01:54Z) - Exploring Precision and Recall to assess the quality and diversity of LLMs [82.21278402856079]
We introduce a novel evaluation framework for Large Language Models (LLMs) such as textscLlama-2 and textscMistral.
This approach allows for a nuanced assessment of the quality and diversity of generated text without the need for aligned corpora.
arXiv Detail & Related papers (2024-02-16T13:53:26Z) - Lost in the Source Language: How Large Language Models Evaluate the Quality of Machine Translation [64.5862977630713]
This study investigates how Large Language Models (LLMs) leverage source and reference data in machine translation evaluation task.
We find that reference information significantly enhances the evaluation accuracy, while surprisingly, source information sometimes is counterproductive.
arXiv Detail & Related papers (2024-01-12T13:23:21Z) - MISMATCH: Fine-grained Evaluation of Machine-generated Text with
Mismatch Error Types [68.76742370525234]
We propose a new evaluation scheme to model human judgments in 7 NLP tasks, based on the fine-grained mismatches between a pair of texts.
Inspired by the recent efforts in several NLP tasks for fine-grained evaluation, we introduce a set of 13 mismatch error types.
We show that the mismatch errors between the sentence pairs on the held-out datasets from 7 NLP tasks align well with the human evaluation.
arXiv Detail & Related papers (2023-06-18T01:38:53Z) - Transfer Learning for Low-Resource Sentiment Analysis [1.2891210250935146]
In this paper, the collection and annotation of a dataset are described for sentiment analysis of Central Kurdish.
We explore a few classical machine learning and neural network-based techniques for this task.
arXiv Detail & Related papers (2023-04-10T16:44:44Z) - HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - Competency-Aware Neural Machine Translation: Can Machine Translation
Know its Own Translation Quality? [61.866103154161884]
Neural machine translation (NMT) is often criticized for failures that happen without awareness.
We propose a novel competency-aware NMT by extending conventional NMT with a self-estimator.
We show that the proposed method delivers outstanding performance on quality estimation.
arXiv Detail & Related papers (2022-11-25T02:39:41Z) - Quality Estimation without Human-labeled Data [25.25993509174361]
Quality estimation aims to measure the quality of translated content without access to a reference translation.
We propose a technique that does not rely on examples from human-annotators and instead uses synthetic training data.
We train off-the-shelf architectures for supervised quality estimation on our synthetic data and show that the resulting models achieve comparable performance to models trained on human-annotated data.
arXiv Detail & Related papers (2021-02-08T06:25:46Z) - Aspects of Terminological and Named Entity Knowledge within Rule-Based
Machine Translation Models for Under-Resourced Neural Machine Translation
Scenarios [3.413805964168321]
Rule-based machine translation is a machine translation paradigm where linguistic knowledge is encoded by an expert.
We describe different approaches to leverage the information contained in rule-based machine translation systems to improve a neural machine translation model.
Our results suggest that the proposed models have limited ability to learn from external information.
arXiv Detail & Related papers (2020-09-28T15:19:23Z) - Computer Assisted Translation with Neural Quality Estimation and
Automatic Post-Editing [18.192546537421673]
We propose an end-to-end deep learning framework of the quality estimation and automatic post-editing of the machine translation output.
Our goal is to provide error correction suggestions and to further relieve the burden of human translators through an interpretable model.
arXiv Detail & Related papers (2020-09-19T00:29:00Z) - Salience Estimation with Multi-Attention Learning for Abstractive Text
Summarization [86.45110800123216]
In the task of text summarization, salience estimation for words, phrases or sentences is a critical component.
We propose a Multi-Attention Learning framework which contains two new attention learning components for salience estimation.
arXiv Detail & Related papers (2020-04-07T02:38:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.