Advancing Precise Outline-Conditioned Text Generation with Task Duality
and Explicit Outline Control
- URL: http://arxiv.org/abs/2305.14459v3
- Date: Wed, 7 Feb 2024 06:28:44 GMT
- Title: Advancing Precise Outline-Conditioned Text Generation with Task Duality
and Explicit Outline Control
- Authors: Yunzhe Li, Qian Chen, Weixiang Yan, Wen Wang, Qinglin Zhang, Hari
Sundaram
- Abstract summary: We introduce a novel text generation task called Precise Outline-conditioned Generation.
This task requires generating stories based on specific, sentence-level outlines.
We propose an explicit outline utilization control approach and a novel framework that leverages the task duality between summarization and generation.
- Score: 15.881568820009797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing works on outline-conditioned text generation typically aim to
generate text using provided outlines as rough sketches, such as keywords and
phrases. However, these approaches make it challenging to control the quality
of text generation and assess consistency between outlines and generated texts
due to lack of clarity and rationality of the rough outlines. In this paper, we
introduce a novel text generation task called Precise Outline-conditioned
Generation, which requires generating stories based on specific, sentence-level
outlines. To facilitate research on this task, we construct two new datasets,
WPOG and CDM. We provide strong baselines based on fine-tuning models such as
BART and GPT-2, and evaluating zero-shot performance of models such as ChatGPT
and Vicuna. Furthermore, we identify an issue of imbalanced utilization of the
outline information in the precise outline-conditioned generation, which is
ubiquitously observed across fine-tuned models and zero-shot inference models.
To address this issue, we propose an explicit outline utilization control
approach and a novel framework that leverages the task duality between
summarization and generation. Experimental results show that the proposed
approaches effectively alleviate the issue of imbalanced outline utilization
and enhance the quality of precise outline-conditioned text generation for both
fine-tuning and zero-shot settings.
Related papers
- Detecting Document-level Paraphrased Machine Generated Content: Mimicking Human Writing Style and Involving Discourse Features [57.34477506004105]
Machine-generated content poses challenges such as academic plagiarism and the spread of misinformation.
We introduce novel methodologies and datasets to overcome these challenges.
We propose MhBART, an encoder-decoder model designed to emulate human writing style.
We also propose DTransformer, a model that integrates discourse analysis through PDTB preprocessing to encode structural features.
arXiv Detail & Related papers (2024-12-17T08:47:41Z) - Exploring Precision and Recall to assess the quality and diversity of LLMs [82.21278402856079]
We introduce a novel evaluation framework for Large Language Models (LLMs) such as textscLlama-2 and textscMistral.
This approach allows for a nuanced assessment of the quality and diversity of generated text without the need for aligned corpora.
arXiv Detail & Related papers (2024-02-16T13:53:26Z) - Text2Data: Low-Resource Data Generation with Textual Control [100.5970757736845]
Text2Data is a novel approach that utilizes unlabeled data to understand the underlying data distribution.
It undergoes finetuning via a novel constraint optimization-based learning objective that ensures controllability and effectively counteracts catastrophic forgetting.
arXiv Detail & Related papers (2024-02-08T03:41:39Z) - Dynamic Relation Transformer for Contextual Text Block Detection [9.644204545582742]
Contextual Text Block Detection is the task of identifying coherent text blocks within the complexity of natural scenes.
Previous methodologies have treated CTBD as either a visual relation extraction challenge within computer vision or as a sequence modeling problem.
We introduce a new framework that frames CTBD as a graph generation problem.
arXiv Detail & Related papers (2024-01-17T14:17:59Z) - Sequentially Controlled Text Generation [97.22539956688443]
GPT-2 generates sentences that are remarkably human-like, longer documents can ramble and do not follow human-like writing structure.
We study the problem of imposing structure on long-range text.
We develop a sequential controlled text generation pipeline with generation and editing.
arXiv Detail & Related papers (2023-01-05T21:23:51Z) - RSTGen: Imbuing Fine-Grained Interpretable Control into Long-FormText
Generators [26.27412809287025]
RSTGen is a framework that controls the discourse structure, semantics and topics of generated text.
We demonstrate our model's ability to control structural discourse and semantic features of generated text in open generation evaluation.
arXiv Detail & Related papers (2022-05-25T09:06:04Z) - Event Transition Planning for Open-ended Text Generation [55.729259805477376]
Open-ended text generation tasks require models to generate a coherent continuation given limited preceding context.
We propose a novel two-stage method which explicitly arranges the ensuing events in open-ended text generation.
Our approach can be understood as a specially-trained coarse-to-fine algorithm.
arXiv Detail & Related papers (2022-04-20T13:37:51Z) - Data-to-text Generation with Variational Sequential Planning [74.3955521225497]
We consider the task of data-to-text generation, which aims to create textual output from non-linguistic input.
We propose a neural model enhanced with a planning component responsible for organizing high-level information in a coherent and meaningful way.
We infer latent plans sequentially with a structured variational model, while interleaving the steps of planning and generation.
arXiv Detail & Related papers (2022-02-28T13:17:59Z) - A Causal Lens for Controllable Text Generation [36.26478600135344]
This paper proposes to formulate controllable text generation from a principled causal perspective.
A direct advantage of the causal formulation is the use of rich causality tools to mitigate generation biases and improve control.
Experiments show significant superiority of the causal approach over previous conditional models for improved control accuracy and reduced bias.
arXiv Detail & Related papers (2022-01-22T19:31:43Z) - Consistency and Coherency Enhanced Story Generation [35.08911595854691]
We propose a two-stage generation framework to enhance consistency and coherency of generated stories.
The first stage is to organize the story outline which depicts the story plots and events, and the second stage is to expand the outline into a complete story.
In addition, coreference supervision signals are incorporated to reduce coreference errors and improve the coreference consistency.
arXiv Detail & Related papers (2020-10-17T16:40:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.