Lexically Constrained Neural Machine Translation with Levenshtein
Transformer
- URL: http://arxiv.org/abs/2004.12681v1
- Date: Mon, 27 Apr 2020 09:59:27 GMT
- Title: Lexically Constrained Neural Machine Translation with Levenshtein
Transformer
- Authors: Raymond Hendy Susanto, Shamil Chollampatt, and Liling Tan
- Abstract summary: This paper proposes a simple and effective algorithm for incorporating lexical constraints in neural machine translation.
Our method injects terminology constraints at inference time without any impact on decoding speed.
- Score: 8.831954614241234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a simple and effective algorithm for incorporating
lexical constraints in neural machine translation. Previous work either
required re-training existing models with the lexical constraints or
incorporating them during beam search decoding with significantly higher
computational overheads. Leveraging the flexibility and speed of a recently
proposed Levenshtein Transformer model (Gu et al., 2019), our method injects
terminology constraints at inference time without any impact on decoding speed.
Our method does not require any modification to the training procedure and can
be easily applied at runtime with custom dictionaries. Experiments on
English-German WMT datasets show that our approach improves an unconstrained
baseline and previous approaches.
Related papers
- An Analysis of BPE Vocabulary Trimming in Neural Machine Translation [56.383793805299234]
vocabulary trimming is a postprocessing step that replaces rare subwords with their component subwords.
We show that vocabulary trimming fails to improve performance and is even prone to incurring heavy degradation.
arXiv Detail & Related papers (2024-03-30T15:29:49Z) - Fast Training of NMT Model with Data Sorting [0.0]
The Transformer model has revolutionized Natural Language Processing tasks such as Neural Machine Translation.
One potential area for improvement is to address the study of empty tokens that the Transformer computes only to discard them later.
We propose an algorithm that sorts sentence pairs based on their length before translation, minimizing the waste of computing power.
arXiv Detail & Related papers (2023-08-16T05:48:50Z) - Downstream Task-Oriented Neural Tokenizer Optimization with Vocabulary
Restriction as Post Processing [4.781986758380065]
This paper proposes a method to optimize tokenization for the performance improvement of already trained downstream models.
Our method generates tokenization results attaining lower loss values of a given downstream model on the training data for restricting vocabularies and trains a tokenizer reproducing the tokenization results.
arXiv Detail & Related papers (2023-04-21T08:29:14Z) - Confident Adaptive Language Modeling [95.45272377648773]
CALM is a framework for dynamically allocating different amounts of compute per input and generation timestep.
We demonstrate the efficacy of our framework in reducing compute -- potential speedup of up to $times 3$ -- while provably maintaining high performance.
arXiv Detail & Related papers (2022-07-14T17:00:19Z) - Improving Pre-trained Language Model Fine-tuning with Noise Stability
Regularization [94.4409074435894]
We propose a novel and effective fine-tuning framework, named Layerwise Noise Stability Regularization (LNSR)
Specifically, we propose to inject the standard Gaussian noise and regularize hidden representations of the fine-tuned model.
We demonstrate the advantages of the proposed method over other state-of-the-art algorithms including L2-SP, Mixout and SMART.
arXiv Detail & Related papers (2022-06-12T04:42:49Z) - A Template-based Method for Constrained Neural Machine Translation [100.02590022551718]
We propose a template-based method that can yield results with high translation quality and match accuracy while keeping the decoding speed.
The generation and derivation of the template can be learned through one sequence-to-sequence training framework.
Experimental results show that the proposed template-based methods can outperform several representative baselines in lexically and structurally constrained translation tasks.
arXiv Detail & Related papers (2022-05-23T12:24:34Z) - Controlled Text Generation as Continuous Optimization with Multiple
Constraints [23.71027518888138]
We propose a flexible and modular algorithm for controllable inference from pretrained models.
We make use of Lagrangian multipliers and gradient-descent based techniques to generate the desired text.
We evaluate our approach on controllable machine translation and style transfer with multiple sentence-level attributes.
arXiv Detail & Related papers (2021-08-04T05:25:20Z) - Encouraging Neural Machine Translation to Satisfy Terminology
Constraints [3.3108924994485096]
We present a new approach to encourage neural machine translation to satisfy lexical constraints.
Our method acts at the training step and thereby avoiding the introduction of any extra computational overhead at inference step.
arXiv Detail & Related papers (2021-06-07T15:46:07Z) - NeuroLogic Decoding: (Un)supervised Neural Text Generation with
Predicate Logic Constraints [75.66980495245926]
Conditional text generation often requires lexical constraints, i.e., which words should or shouldn't be included in the output text.
We propose NeuroLogic Decoding, a simple yet effective algorithm that enables neural language models -- supervised or not -- to generate fluent text.
Our results suggest the limit of large-scale neural networks for fine-grained controllable generation and the promise of inference-time algorithms.
arXiv Detail & Related papers (2020-10-24T11:55:22Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.