Disambiguated Lexically Constrained Neural Machine Translation
- URL: http://arxiv.org/abs/2305.17351v1
- Date: Sat, 27 May 2023 03:15:10 GMT
- Title: Disambiguated Lexically Constrained Neural Machine Translation
- Authors: Jinpeng Zhang, Nini Xiao, Ke Wang, Chuanqi Dong, Xiangyu Duan, Yuqi
Zhang, Min Zhang
- Abstract summary: Current approaches to LCNMT assume that the pre-specified lexical constraints are contextually appropriate.
We propose disambiguated LCNMT (D-LCNMT) to solve the problem.
D-LCNMT is a robust and effective two-stage framework that disambiguates the constraints based on contexts at first, then integrates the disambiguated constraints into LCNMT.
- Score: 20.338107081523212
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Lexically constrained neural machine translation (LCNMT), which controls the
translation generation with pre-specified constraints, is important in many
practical applications. Current approaches to LCNMT typically assume that the
pre-specified lexical constraints are contextually appropriate. This assumption
limits their application to real-world scenarios where a source lexicon may
have multiple target constraints, and disambiguation is needed to select the
most suitable one. In this paper, we propose disambiguated LCNMT (D-LCNMT) to
solve the problem. D-LCNMT is a robust and effective two-stage framework that
disambiguates the constraints based on contexts at first, then integrates the
disambiguated constraints into LCNMT. Experimental results show that our
approach outperforms strong baselines including existing data augmentation
based approaches on benchmark datasets, and comprehensive experiments in
scenarios where a source lexicon corresponds to multiple target constraints
demonstrate the constraint disambiguation superiority of our approach.
Related papers
- Attribute Controlled Fine-tuning for Large Language Models: A Case Study on Detoxification [76.14641982122696]
We propose a constraint learning schema for fine-tuning Large Language Models (LLMs) with attribute control.
We show that our approach leads to an LLM that produces fewer inappropriate responses while achieving competitive performance on benchmarks and a toxicity detection task.
arXiv Detail & Related papers (2024-10-07T23:38:58Z) - Making Large Language Models Better Planners with Reasoning-Decision Alignment [70.5381163219608]
We motivate an end-to-end decision-making model based on multimodality-augmented LLM.
We propose a reasoning-decision alignment constraint between the paired CoTs and planning results.
We dub our proposed large language planners with reasoning-decision alignment as RDA-Driver.
arXiv Detail & Related papers (2024-08-25T16:43:47Z) - Translate-and-Revise: Boosting Large Language Models for Constrained Translation [42.37981028583618]
We leverage the capabilities of large language models (LLMs) for constrained translation.
LLMs can easily adapt to this task by taking translation instructions and constraints as prompts.
We show 15% improvement in constraint-based translation accuracy over standard LLMs.
arXiv Detail & Related papers (2024-07-18T05:08:09Z) - Prefix Text as a Yarn: Eliciting Non-English Alignment in Foundation Language Model [50.339632513018934]
supervised fine-tuning (SFT) has been a straightforward approach for tailoring the output of foundation large language model (LLM) to specific preferences.
We critically examine this hypothesis within the scope of cross-lingual generation tasks.
We introduce a novel training-free alignment method named PreTTY, which employs minimal task-related prior tokens.
arXiv Detail & Related papers (2024-04-25T17:19:36Z) - FollowBench: A Multi-level Fine-grained Constraints Following Benchmark for Large Language Models [79.62191017182518]
FollowBench is a benchmark for Fine-grained Constraints Following Benchmark for Large Language Models.
We introduce a Multi-level mechanism that incrementally adds a single constraint to the initial instruction at each increased level.
By evaluating 13 popular LLMs on FollowBench, we highlight the weaknesses of LLMs in instruction following and point towards potential avenues for future work.
arXiv Detail & Related papers (2023-10-31T12:32:38Z) - Towards Accurate Translation via Semantically Appropriate Application of
Lexical Constraints [0.0]
We focus on two important but under-studied issues that lie in the current evaluation process of LNMT studies.
The model needs to cope with challenging lexical constraints that are "homographs" or "unseen" during training.
We propose PLUMCOT, which integrates contextually rich information about unseen lexical constraints from pre-trained language models.
We also release HOLLY, an evaluation benchmark for assessing the ability of a model to cope with "homographic" and "unseen" lexical constraints.
arXiv Detail & Related papers (2023-06-21T08:08:15Z) - Neural Fields with Hard Constraints of Arbitrary Differential Order [61.49418682745144]
We develop a series of approaches for enforcing hard constraints on neural fields.
The constraints can be specified as a linear operator applied to the neural field and its derivatives.
Our approaches are demonstrated in a wide range of real-world applications.
arXiv Detail & Related papers (2023-06-15T08:33:52Z) - Integrating Vectorized Lexical Constraints for Neural Machine
Translation [22.300632179228426]
Lexically constrained neural machine translation (NMT) controls the generation of NMT models with pre-specified constraints.
Most existing works choose to construct synthetic data or modify the decoding algorithm to impose lexical constraints, treating the NMT model as a black box.
We propose to open this black box by directly integrating the constraints into NMT models.
arXiv Detail & Related papers (2022-03-23T05:54:37Z) - Rule-based Morphological Inflection Improves Neural Terminology
Translation [16.802947102163497]
We introduce a modular framework for incorporating lemma constraints in neural MT (NMT)
It is based on a novel cross-lingual inflection module that inflects the target lemma constraints based on the source context.
Results show that our rule-based inflection module helps NMT models incorporate lemma constraints more accurately than a neural module and outperforms the existing end-to-end approach with lower training costs.
arXiv Detail & Related papers (2021-09-10T02:06:48Z) - NeuroLogic Decoding: (Un)supervised Neural Text Generation with
Predicate Logic Constraints [75.66980495245926]
Conditional text generation often requires lexical constraints, i.e., which words should or shouldn't be included in the output text.
We propose NeuroLogic Decoding, a simple yet effective algorithm that enables neural language models -- supervised or not -- to generate fluent text.
Our results suggest the limit of large-scale neural networks for fine-grained controllable generation and the promise of inference-time algorithms.
arXiv Detail & Related papers (2020-10-24T11:55:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.