Look It Up: Bilingual Dictionaries Improve Neural Machine Translation
- URL: http://arxiv.org/abs/2010.05997v2
- Date: Fri, 28 Jan 2022 17:09:35 GMT
- Title: Look It Up: Bilingual Dictionaries Improve Neural Machine Translation
- Authors: Xing Jie Zhong, and David Chiang
- Abstract summary: We describe a new method for "attaching" dictionary definitions to rare words so that the network can learn the best way to use them.
We demonstrate improvements of up to 1.8 BLEU using bilingual dictionaries.
- Score: 17.385945558427863
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite advances in neural machine translation (NMT) quality, rare words
continue to be problematic. For humans, the solution to the rare-word problem
has long been dictionaries, but dictionaries cannot be straightforwardly
incorporated into NMT. In this paper, we describe a new method for "attaching"
dictionary definitions to rare words so that the network can learn the best way
to use them. We demonstrate improvements of up to 1.8 BLEU using bilingual
dictionaries.
Related papers
- Improving Rare Word Translation With Dictionaries and Attention Masking [8.908747084128397]
We propose appending definitions from a bilingual dictionary to source sentences and using attention masking to link together rare words with their definitions.
We find that including definitions for rare words improves performance by up to 1.0 BLEU and 1.6 MacroF1.
arXiv Detail & Related papers (2024-08-17T02:26:29Z) - Towards Effective Disambiguation for Machine Translation with Large
Language Models [65.80775710657672]
We study the capabilities of large language models to translate "ambiguous sentences"
Experiments show that our methods can match or outperform state-of-the-art systems such as DeepL and NLLB in four out of five language directions.
arXiv Detail & Related papers (2023-09-20T22:22:52Z) - DICTDIS: Dictionary Constrained Disambiguation for Improved NMT [50.888881348723295]
We present DictDis, a lexically constrained NMT system that disambiguates between multiple candidate translations derived from dictionaries.
We demonstrate the utility of DictDis via extensive experiments on English-Hindi and English-German sentences in a variety of domains including regulatory, finance, engineering.
arXiv Detail & Related papers (2022-10-13T13:04:16Z) - Dict-NMT: Bilingual Dictionary based NMT for Extremely Low Resource
Languages [1.8787713898828164]
We present a detailed analysis of the effects of the quality of dictionaries, training dataset size, language family, etc., on the translation quality.
Results on multiple low-resource test languages show a clear advantage of our bilingual dictionary-based method over the baselines.
arXiv Detail & Related papers (2022-06-09T12:03:29Z) - When do Contrastive Word Alignments Improve Many-to-many Neural Machine
Translation? [33.28706502928905]
This work proposes a word-level contrastive objective to leverage word alignments for many-to-many NMT.
Analyses reveal that in many-to-many NMT, the encoder's sentence retrieval performance highly correlates with the translation quality.
arXiv Detail & Related papers (2022-04-26T09:07:51Z) - BitextEdit: Automatic Bitext Editing for Improved Low-Resource Machine
Translation [53.55009917938002]
We propose to refine the mined bitexts via automatic editing.
Experiments demonstrate that our approach successfully improves the quality of CCMatrix mined bitext for 5 low-resource language-pairs and 10 translation directions by up to 8 BLEU points.
arXiv Detail & Related papers (2021-11-12T16:00:39Z) - Lacking the embedding of a word? Look it up into a traditional
dictionary [0.2624902795082451]
We propose to use definitions retrieved in traditional dictionaries to produce word embeddings for rare words.
DefiNNet and DefBERT significantly outperform state-of-the-art as well as baseline methods for producing embeddings of unknown words.
arXiv Detail & Related papers (2021-09-24T06:27:58Z) - Phrase-level Active Learning for Neural Machine Translation [107.28450614074002]
We propose an active learning setting where we can spend a given budget on translating in-domain data.
We select both full sentences and individual phrases from unlabelled data in the new domain for routing to human translators.
In a German-English translation task, our active learning approach achieves consistent improvements over uncertainty-based sentence selection methods.
arXiv Detail & Related papers (2021-06-21T19:20:42Z) - Continuous Learning in Neural Machine Translation using Bilingual
Dictionaries [14.058642647656301]
We propose an evaluation framework to assess the ability of neural machine translation to continuously learn new phrases.
By addressing both challenges we are able to improve the ability to translate new, rare words and phrases from 30% to up to 70%.
arXiv Detail & Related papers (2021-02-12T14:46:13Z) - Bilingual Dictionary Based Neural Machine Translation without Using
Parallel Sentences [45.99290614777277]
We propose a new task of machine translation (MT) based on no parallel sentences but can refer to a ground-truth bilingual dictionary.
Motivated by the ability of a monolingual speaker learning to translate via looking up the bilingual dictionary, we propose the task to see how much potential an MT system can attain.
arXiv Detail & Related papers (2020-07-06T12:05:27Z) - When Dictionary Learning Meets Deep Learning: Deep Dictionary Learning
and Coding Network for Image Recognition with Limited Data [74.75557280245643]
We present a new Deep Dictionary Learning and Coding Network (DDLCN) for image recognition tasks with limited data.
We empirically compare DDLCN with several leading dictionary learning methods and deep learning models.
Experimental results on five popular datasets show that DDLCN achieves competitive results compared with state-of-the-art methods when the training data is limited.
arXiv Detail & Related papers (2020-05-21T23:12:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.