Lexically Cohesive Neural Machine Translation with Copy Mechanism
- URL: http://arxiv.org/abs/2010.05193v1
- Date: Sun, 11 Oct 2020 08:39:02 GMT
- Title: Lexically Cohesive Neural Machine Translation with Copy Mechanism
- Authors: Vipul Mishra, Chenhui Chu and Yuki Arase
- Abstract summary: We employ a copy mechanism into a context-aware neural machine translation model to allow copying words from previous outputs.
We conduct experiments on Japanese to English translation using an evaluation dataset for discourse translation.
- Score: 21.43163704217968
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lexically cohesive translations preserve consistency in word choices in
document-level translation. We employ a copy mechanism into a context-aware
neural machine translation model to allow copying words from previous
translation outputs. Different from previous context-aware neural machine
translation models that handle all the discourse phenomena implicitly, our
model explicitly addresses the lexical cohesion problem by boosting the
probabilities to output words consistently. We conduct experiments on Japanese
to English translation using an evaluation dataset for discourse translation.
The results showed that the proposed model significantly improved lexical
cohesion compared to previous context-aware models.
Related papers
- Is Robustness Transferable across Languages in Multilingual Neural
Machine Translation? [45.04661608619081]
We investigate the transferability of robustness across different languages in multilingual neural machine translation.
Our findings demonstrate that the robustness gained in one translation direction can indeed transfer to other translation directions.
arXiv Detail & Related papers (2023-10-31T04:10:31Z) - Decomposed Prompting for Machine Translation Between Related Languages
using Large Language Models [55.35106713257871]
We introduce DecoMT, a novel approach of few-shot prompting that decomposes the translation process into a sequence of word chunk translations.
We show that DecoMT outperforms the strong few-shot prompting BLOOM model with an average improvement of 8 chrF++ scores across the examined languages.
arXiv Detail & Related papers (2023-05-22T14:52:47Z) - HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - As Little as Possible, as Much as Necessary: Detecting Over- and
Undertranslations with Contrastive Conditioning [42.46681912294797]
We propose a method for detecting superfluous words in neural machine translation.
We compare the likelihood of a full sequence under a translation model to the likelihood of its parts, given the corresponding source or target sequence.
This allows to pinpoint superfluous words in the translation and untranslated words in the source even in the absence of a reference translation.
arXiv Detail & Related papers (2022-03-03T18:59:02Z) - DEEP: DEnoising Entity Pre-training for Neural Machine Translation [123.6686940355937]
It has been shown that machine translation models usually generate poor translations for named entities that are infrequent in the training corpus.
We propose DEEP, a DEnoising Entity Pre-training method that leverages large amounts of monolingual data and a knowledge base to improve named entity translation accuracy within sentences.
arXiv Detail & Related papers (2021-11-14T17:28:09Z) - End-to-End Lexically Constrained Machine Translation for Morphologically
Rich Languages [0.0]
We investigate mechanisms to allow neural machine translation to infer the correct word inflection given lemmatized constraints.
Our experiments on the English-Czech language pair show that this approach improves the translation of constrained terms in both automatic and manual evaluation.
arXiv Detail & Related papers (2021-06-23T13:40:13Z) - Exemplar-Controllable Paraphrasing and Translation using Bitext [57.92051459102902]
We adapt models from prior work to be able to learn solely from bilingual text (bitext)
Our single proposed model can perform four tasks: controlled paraphrase generation in both languages and controlled machine translation in both language directions.
arXiv Detail & Related papers (2020-10-12T17:02:50Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z) - Translation Artifacts in Cross-lingual Transfer Learning [51.66536640084888]
We show that machine translation can introduce subtle artifacts that have a notable impact in existing cross-lingual models.
In natural language inference, translating the premise and the hypothesis independently can reduce the lexical overlap between them.
We also improve the state-of-the-art in XNLI for the translate-test and zero-shot approaches by 4.3 and 2.8 points, respectively.
arXiv Detail & Related papers (2020-04-09T17:54:30Z) - Incorporating Bilingual Dictionaries for Low Resource Semi-Supervised
Neural Machine Translation [5.958653653305609]
We incorporate widely available bilingual dictionaries that yield word-by-word translations to generate synthetic sentences.
This automatically expands the vocabulary of the model while maintaining high quality content.
arXiv Detail & Related papers (2020-04-05T02:14:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.