Advancing Precise Outline-Conditioned Text Generation with Task Duality
and Explicit Outline Control
- URL: http://arxiv.org/abs/2305.14459v3
- Date: Wed, 7 Feb 2024 06:28:44 GMT
- Title: Advancing Precise Outline-Conditioned Text Generation with Task Duality
and Explicit Outline Control
- Authors: Yunzhe Li, Qian Chen, Weixiang Yan, Wen Wang, Qinglin Zhang, Hari
Sundaram
- Abstract summary: We introduce a novel text generation task called Precise Outline-conditioned Generation.
This task requires generating stories based on specific, sentence-level outlines.
We propose an explicit outline utilization control approach and a novel framework that leverages the task duality between summarization and generation.
- Score: 15.881568820009797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing works on outline-conditioned text generation typically aim to
generate text using provided outlines as rough sketches, such as keywords and
phrases. However, these approaches make it challenging to control the quality
of text generation and assess consistency between outlines and generated texts
due to lack of clarity and rationality of the rough outlines. In this paper, we
introduce a novel text generation task called Precise Outline-conditioned
Generation, which requires generating stories based on specific, sentence-level
outlines. To facilitate research on this task, we construct two new datasets,
WPOG and CDM. We provide strong baselines based on fine-tuning models such as
BART and GPT-2, and evaluating zero-shot performance of models such as ChatGPT
and Vicuna. Furthermore, we identify an issue of imbalanced utilization of the
outline information in the precise outline-conditioned generation, which is
ubiquitously observed across fine-tuned models and zero-shot inference models.
To address this issue, we propose an explicit outline utilization control
approach and a novel framework that leverages the task duality between
summarization and generation. Experimental results show that the proposed
approaches effectively alleviate the issue of imbalanced outline utilization
and enhance the quality of precise outline-conditioned text generation for both
fine-tuning and zero-shot settings.
Related papers
- Contextualized Diffusion Models for Text-Guided Image and Video Generation [67.69171154637172]
Conditional diffusion models have exhibited superior performance in high-fidelity text-guided visual generation and editing.
We propose a novel and general contextualized diffusion model (ContextDiff) by incorporating the cross-modal context encompassing interactions and alignments between text condition and visual sample.
We generalize our model to both DDPMs and DDIMs with theoretical derivations, and demonstrate the effectiveness of our model in evaluations with two challenging tasks: text-to-image generation, and text-to-video editing.
arXiv Detail & Related papers (2024-02-26T15:01:16Z) - Exploring Precision and Recall to assess the quality and diversity of LLMs [82.21278402856079]
We introduce a novel evaluation framework for Large Language Models (LLMs) such as textscLlama-2 and textscMistral.
This approach allows for a nuanced assessment of the quality and diversity of generated text without the need for aligned corpora.
arXiv Detail & Related papers (2024-02-16T13:53:26Z) - Text2Data: Low-Resource Data Generation with Textual Control [104.38011760992637]
Natural language serves as a common and straightforward control signal for humans to interact seamlessly with machines.
We propose Text2Data, a novel approach that utilizes unlabeled data to understand the underlying data distribution through an unsupervised diffusion model.
It undergoes controllable finetuning via a novel constraint optimization-based learning objective that ensures controllability and effectively counteracts catastrophic forgetting.
arXiv Detail & Related papers (2024-02-08T03:41:39Z) - Dynamic Relation Transformer for Contextual Text Block Detection [9.644204545582742]
Contextual Text Block Detection is the task of identifying coherent text blocks within the complexity of natural scenes.
Previous methodologies have treated CTBD as either a visual relation extraction challenge within computer vision or as a sequence modeling problem.
We introduce a new framework that frames CTBD as a graph generation problem.
arXiv Detail & Related papers (2024-01-17T14:17:59Z) - Sequentially Controlled Text Generation [97.22539956688443]
GPT-2 generates sentences that are remarkably human-like, longer documents can ramble and do not follow human-like writing structure.
We study the problem of imposing structure on long-range text.
We develop a sequential controlled text generation pipeline with generation and editing.
arXiv Detail & Related papers (2023-01-05T21:23:51Z) - RSTGen: Imbuing Fine-Grained Interpretable Control into Long-FormText
Generators [26.27412809287025]
RSTGen is a framework that controls the discourse structure, semantics and topics of generated text.
We demonstrate our model's ability to control structural discourse and semantic features of generated text in open generation evaluation.
arXiv Detail & Related papers (2022-05-25T09:06:04Z) - Event Transition Planning for Open-ended Text Generation [55.729259805477376]
Open-ended text generation tasks require models to generate a coherent continuation given limited preceding context.
We propose a novel two-stage method which explicitly arranges the ensuing events in open-ended text generation.
Our approach can be understood as a specially-trained coarse-to-fine algorithm.
arXiv Detail & Related papers (2022-04-20T13:37:51Z) - Data-to-text Generation with Variational Sequential Planning [74.3955521225497]
We consider the task of data-to-text generation, which aims to create textual output from non-linguistic input.
We propose a neural model enhanced with a planning component responsible for organizing high-level information in a coherent and meaningful way.
We infer latent plans sequentially with a structured variational model, while interleaving the steps of planning and generation.
arXiv Detail & Related papers (2022-02-28T13:17:59Z) - A Causal Lens for Controllable Text Generation [36.26478600135344]
This paper proposes to formulate controllable text generation from a principled causal perspective.
A direct advantage of the causal formulation is the use of rich causality tools to mitigate generation biases and improve control.
Experiments show significant superiority of the causal approach over previous conditional models for improved control accuracy and reduced bias.
arXiv Detail & Related papers (2022-01-22T19:31:43Z) - Consistency and Coherency Enhanced Story Generation [35.08911595854691]
We propose a two-stage generation framework to enhance consistency and coherency of generated stories.
The first stage is to organize the story outline which depicts the story plots and events, and the second stage is to expand the outline into a complete story.
In addition, coreference supervision signals are incorporated to reduce coreference errors and improve the coreference consistency.
arXiv Detail & Related papers (2020-10-17T16:40:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.