Extract, Denoise, and Enforce: Evaluating and Predicting Lexical
Constraints for Conditional Text Generation
- URL: http://arxiv.org/abs/2104.08724v1
- Date: Sun, 18 Apr 2021 05:29:02 GMT
- Title: Extract, Denoise, and Enforce: Evaluating and Predicting Lexical
Constraints for Conditional Text Generation
- Authors: Yuning Mao, Wenchang Ma, Deren Lei, Xiang Ren
- Abstract summary: We present a systematic analysis of conditional generation to study whether current PLMs are good enough for preserving important concepts in the input.
We propose a framework for automatic constraint extraction, denoising, and enforcement that is shown to perform comparably or better than unconstrained generation.
- Score: 31.341566859483056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, pre-trained language models (PLMs) have dominated conditional text
generation tasks. Given the impressive performance and prevalence of the PLMs,
it is seemingly natural to assume that they could figure out what to attend to
in the input and what to include in the output via seq2seq learning without
more guidance than the training input/output pairs. However, a rigorous study
regarding the above assumption is still lacking. In this paper, we present a
systematic analysis of conditional generation to study whether current PLMs are
good enough for preserving important concepts in the input and to what extent
explicitly guiding generation with lexical constraints is beneficial. We
conduct extensive analytical experiments on a range of conditional generation
tasks and try to answer in what scenarios guiding generation with lexical
constraints works well and why. We then propose a framework for automatic
constraint extraction, denoising, and enforcement that is shown to perform
comparably or better than unconstrained generation. We hope that our findings
could serve as a reference when determining whether it is appropriate and
worthwhile to use explicit constraints for a specific task or
dataset.\footnote{Our code is available at
\url{https://github.com/morningmoni/LCGen-eval}.}
Related papers
- Graph-DPEP: Decomposed Plug and Ensemble Play for Few-Shot Document Relation Extraction with Graph-of-Thoughts Reasoning [34.85741925091139]
Graph-DPEP framework is grounded in the reasoning behind triplet explanation thoughts presented in natural language.
We develop "ensemble-play", reapplying generation on the entire type list by leveraging the reasoning thoughts embedded in a sub-graph.
arXiv Detail & Related papers (2024-11-05T07:12:36Z) - Unlocking Anticipatory Text Generation: A Constrained Approach for Large Language Models Decoding [75.06872859716049]
Large Language Models (LLMs) have demonstrated a powerful ability for text generation.
undesired behaviors such as toxicity or hallucinations can manifest.
We propose formalizing text generation as a future-constrained generation problem.
arXiv Detail & Related papers (2023-12-11T06:35:33Z) - Evaluating, Understanding, and Improving Constrained Text Generation for Large Language Models [49.74036826946397]
This study investigates constrained text generation for large language models (LLMs)
Our research mainly focuses on mainstream open-source LLMs, categorizing constraints into lexical, structural, and relation-based types.
Results illuminate LLMs' capacity and deficiency to incorporate constraints and provide insights for future developments in constrained text generation.
arXiv Detail & Related papers (2023-10-25T03:58:49Z) - Towards Accurate Translation via Semantically Appropriate Application of
Lexical Constraints [0.0]
We focus on two important but under-studied issues that lie in the current evaluation process of LNMT studies.
The model needs to cope with challenging lexical constraints that are "homographs" or "unseen" during training.
We propose PLUMCOT, which integrates contextually rich information about unseen lexical constraints from pre-trained language models.
We also release HOLLY, an evaluation benchmark for assessing the ability of a model to cope with "homographic" and "unseen" lexical constraints.
arXiv Detail & Related papers (2023-06-21T08:08:15Z) - Lexically-constrained Text Generation through Commonsense Knowledge
Extraction and Injection [62.071938098215085]
We focus on the Commongen benchmark, wherein the aim is to generate a plausible sentence for a given set of input concepts.
We propose strategies for enhancing the semantic correctness of the generated text.
arXiv Detail & Related papers (2020-12-19T23:23:40Z) - NeuroLogic Decoding: (Un)supervised Neural Text Generation with
Predicate Logic Constraints [75.66980495245926]
Conditional text generation often requires lexical constraints, i.e., which words should or shouldn't be included in the output text.
We propose NeuroLogic Decoding, a simple yet effective algorithm that enables neural language models -- supervised or not -- to generate fluent text.
Our results suggest the limit of large-scale neural networks for fine-grained controllable generation and the promise of inference-time algorithms.
arXiv Detail & Related papers (2020-10-24T11:55:22Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z) - ColdGANs: Taming Language GANs with Cautious Sampling Strategies [29.943949944682196]
Generative Adversarial Networks (GANs) can mitigate limitations but the discrete nature of text has hindered their application to language generation.
We show how classical sampling results in unstable training.
We propose to consider alternative exploration strategies in a GAN framework that we name ColdGANs, where we force the sampling to be close to the distribution modes to get smoother learning dynamics.
For the first time, to the best of our knowledge, the proposed language GANs compare favorably to MLE, and obtain improvements over the state-of-the-art on three generative tasks.
arXiv Detail & Related papers (2020-06-08T14:48:14Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.