MDQE: A More Accurate Direct Pretraining for Machine Translation Quality
Estimation
- URL: http://arxiv.org/abs/2107.14600v1
- Date: Sat, 24 Jul 2021 09:48:37 GMT
- Title: MDQE: A More Accurate Direct Pretraining for Machine Translation Quality
Estimation
- Authors: Lei Lin
- Abstract summary: We argue that there are still gaps between the predictor and the estimator in both data quality and training objectives.
We propose a novel framework that provides a more accurate direct pretraining for QE tasks.
- Score: 4.416484585765028
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: It is expensive to evaluate the results of Machine Translation(MT), which
usually requires manual translation as a reference. Machine Translation Quality
Estimation (QE) is a task of predicting the quality of machine translations
without relying on any reference. Recently, the emergence of
predictor-estimator framework which trains the predictor as a feature extractor
and estimator as a QE predictor, and pre-trained language models(PLM) have
achieved promising QE performance. However, we argue that there are still gaps
between the predictor and the estimator in both data quality and training
objectives, which preclude QE models from benefiting from a large number of
parallel corpora more directly. Based on previous related work that have
alleviated gaps to some extent, we propose a novel framework that provides a
more accurate direct pretraining for QE tasks. In this framework, a generator
is trained to produce pseudo data that is closer to the real QE data, and a
estimator is pretrained on these data with novel objectives that are the same
as the QE task. Experiments on widely used benchmarks show that our proposed
framework outperforms existing methods, without using any pretraining models
such as BERT.
Related papers
- Improving Machine Translation with Human Feedback: An Exploration of Quality Estimation as a Reward Model [75.66013048128302]
In this work, we investigate the potential of employing the QE model as the reward model to predict human preferences for feedback training.
We first identify the overoptimization problem during QE-based feedback training, manifested as an increase in reward while translation quality declines.
To address the problem, we adopt a simple yet effective method that uses rules to detect the incorrect translations and assigns a penalty term to the reward scores of them.
arXiv Detail & Related papers (2024-01-23T16:07:43Z) - How to Estimate Model Transferability of Pre-Trained Speech Models? [84.11085139766108]
"Score-based assessment" framework for estimating transferability of pre-trained speech models.
We leverage upon two representation theories, Bayesian likelihood estimation and optimal transport, to generate rank scores for the PSM candidates.
Our framework efficiently computes transferability scores without actual fine-tuning of candidate models or layers.
arXiv Detail & Related papers (2023-06-01T04:52:26Z) - PreQuEL: Quality Estimation of Machine Translation Outputs in Advance [32.922128367314194]
A PreQuEL system predicts how well a given sentence will be translated, without recourse to the actual translation.
We develop a baseline model for the task and analyze its performance.
We show that this augmentation method can improve the performance of the Quality-Estimation task as well.
arXiv Detail & Related papers (2022-05-18T18:55:05Z) - Translation Error Detection as Rationale Extraction [36.616561917049076]
We study the behaviour of state-of-the-art sentence-level QE models and show that explanations can indeed be used to detect translation errors.
We introduce a novel semi-supervised method for word-level QE and (ii) propose to use the QE task as a new benchmark for evaluating the plausibility of feature attribution.
arXiv Detail & Related papers (2021-08-27T09:35:14Z) - Knowledge Distillation for Quality Estimation [79.51452598302934]
Quality Estimation (QE) is the task of automatically predicting Machine Translation quality in the absence of reference translations.
Recent success in QE stems from the use of multilingual pre-trained representations, where very large models lead to impressive results.
We show that this approach, in combination with data augmentation, leads to light-weight QE models that perform competitively with distilled pre-trained representations with 8x fewer parameters.
arXiv Detail & Related papers (2021-07-01T12:36:21Z) - Ensemble-based Transfer Learning for Low-resource Machine Translation
Quality Estimation [1.7188280334580195]
We focus on the Sentence-Level QE Shared Task of the Fifth Conference on Machine Translation (WMT20)
We propose an ensemble-based predictor-estimator QE model with transfer learning to overcome such QE data scarcity challenge.
We achieve the best performance on the ensemble model combining the models pretrained by individual languages as well as different levels of parallel trained corpus with a Pearson's correlation of 0.298.
arXiv Detail & Related papers (2021-05-17T06:02:17Z) - DirectQE: Direct Pretraining for Machine Translation Quality Estimation [41.187833219223336]
We argue that there are gaps between the predictor and the estimator in both data quality and training objectives.
We propose a novel framework called DirectQE that provides a direct pretraining for QE tasks.
arXiv Detail & Related papers (2021-05-15T06:18:49Z) - Unsupervised Quality Estimation for Neural Machine Translation [63.38918378182266]
Existing approaches require large amounts of expert annotated data, computation and time for training.
We devise an unsupervised approach to QE where no training or access to additional resources besides the MT system itself is required.
We achieve very good correlation with human judgments of quality, rivalling state-of-the-art supervised QE models.
arXiv Detail & Related papers (2020-05-21T12:38:06Z) - Pre-training Is (Almost) All You Need: An Application to Commonsense
Reasoning [61.32992639292889]
Fine-tuning of pre-trained transformer models has become the standard approach for solving common NLP tasks.
We introduce a new scoring method that casts a plausibility ranking task in a full-text format.
We show that our method provides a much more stable training phase across random restarts.
arXiv Detail & Related papers (2020-04-29T10:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.