PoetryDiffusion: Towards Joint Semantic and Metrical Manipulation in
Poetry Generation
- URL: http://arxiv.org/abs/2306.08456v3
- Date: Tue, 19 Dec 2023 17:42:02 GMT
- Title: PoetryDiffusion: Towards Joint Semantic and Metrical Manipulation in
Poetry Generation
- Authors: Zhiyuan Hu, Chumin Liu, Yue Feng, Anh Tuan Luu, Bryan Hooi
- Abstract summary: Controllable text generation is a challenging and meaningful field in natural language generation (NLG)
In this paper, we pioneer the use of the Diffusion model for generating sonnets and Chinese SongCi poetry.
Our model outperforms existing models in automatic evaluation of semantic, metrical, and overall performance as well as human evaluation.
- Score: 58.36105306993046
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Controllable text generation is a challenging and meaningful field in natural
language generation (NLG). Especially, poetry generation is a typical one with
well-defined and strict conditions for text generation which is an ideal
playground for the assessment of current methodologies. While prior works
succeeded in controlling either semantic or metrical aspects of poetry
generation, simultaneously addressing both remains a challenge. In this paper,
we pioneer the use of the Diffusion model for generating sonnets and Chinese
SongCi poetry to tackle such challenges. In terms of semantics, our
PoetryDiffusion model, built upon the Diffusion model, generates entire
sentences or poetry by comprehensively considering the entirety of sentence
information. This approach enhances semantic expression, distinguishing it from
autoregressive and large language models (LLMs). For metrical control, the
separation feature of diffusion generation and its constraint control module
enable us to flexibly incorporate a novel metrical controller to manipulate and
evaluate metrics (format and rhythm). The denoising process in PoetryDiffusion
allows for gradual enhancement of semantics and flexible integration of the
metrical controller which can calculate and impose penalties on states that
stray significantly from the target control distribution. Experimental results
on two datasets demonstrate that our model outperforms existing models in
automatic evaluation of semantic, metrical, and overall performance as well as
human evaluation.
Related papers
- GPT Czech Poet: Generation of Czech Poetic Strophes with Language Models [0.4444634303550442]
We introduce a new model for generating poetry in Czech language, based on fine-tuning a pre-trained Large Language Model.
We demonstrate that guiding the generation process by explicitly specifying strophe parameters within the poem text strongly improves the effectiveness of the model.
arXiv Detail & Related papers (2024-06-18T06:19:45Z) - Controllable Generation with Text-to-Image Diffusion Models: A Survey [8.394970202694529]
controllable generation studies aim to control pre-trained text-to-image (T2I) models to support novel conditions.
Our review begins with a brief introduction to the basics of denoising diffusion probabilistic models.
We then reveal the controlling mechanisms of diffusion models, theoretically analyzing how novel conditions are introduced into the denoising process.
arXiv Detail & Related papers (2024-03-07T07:24:18Z) - Contextualized Diffusion Models for Text-Guided Image and Video Generation [67.69171154637172]
Conditional diffusion models have exhibited superior performance in high-fidelity text-guided visual generation and editing.
We propose a novel and general contextualized diffusion model (ContextDiff) by incorporating the cross-modal context encompassing interactions and alignments between text condition and visual sample.
We generalize our model to both DDPMs and DDIMs with theoretical derivations, and demonstrate the effectiveness of our model in evaluations with two challenging tasks: text-to-image generation, and text-to-video editing.
arXiv Detail & Related papers (2024-02-26T15:01:16Z) - Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution [67.9215891673174]
We propose score entropy as a novel loss that naturally extends score matching to discrete spaces.
We test our Score Entropy Discrete Diffusion models on standard language modeling tasks.
arXiv Detail & Related papers (2023-10-25T17:59:12Z) - Language Model Decoding as Direct Metrics Optimization [87.68281625776282]
Current decoding methods struggle to generate texts that align with human texts across different aspects.
In this work, we frame decoding from a language model as an optimization problem with the goal of strictly matching the expected performance with human texts.
We prove that this induced distribution is guaranteed to improve the perplexity on human texts, which suggests a better approximation to the underlying distribution of human texts.
arXiv Detail & Related papers (2023-10-02T09:35:27Z) - PLANNER: Generating Diversified Paragraph via Latent Language Diffusion Model [37.2192243883707]
We propose PLANNER, a model that combines latent semantic diffusion with autoregressive generation to generate fluent text.
Results on semantic generation, text completion and summarization show its effectiveness in generating high-quality long-form text.
arXiv Detail & Related papers (2023-06-05T01:36:39Z) - TESS: Text-to-Text Self-Conditioned Simplex Diffusion [56.881170312435444]
Text-to-text Self-conditioned Simplex Diffusion employs a new form of self-conditioning, and applies the diffusion process on the logit simplex space rather than the learned embedding space.
We demonstrate that TESS outperforms state-of-the-art non-autoregressive models, requires fewer diffusion steps with minimal drop in performance, and is competitive with pretrained autoregressive sequence-to-sequence models.
arXiv Detail & Related papers (2023-05-15T06:33:45Z) - A Cheaper and Better Diffusion Language Model with Soft-Masked Noise [62.719656543880596]
Masked-Diffuse LM is a novel diffusion model for language modeling, inspired by linguistic features in languages.
Specifically, we design a linguistic-informed forward process which adds corruptions to the text through strategically soft-masking to better noise the textual data.
We demonstrate that our Masked-Diffuse LM can achieve better generation quality than the state-of-the-art diffusion models with better efficiency.
arXiv Detail & Related papers (2023-04-10T17:58:42Z) - Self-conditioned Embedding Diffusion for Text Generation [28.342735885752493]
Self-conditioned Embedding Diffusion is a continuous diffusion mechanism that operates on token embeddings.
We show that our text diffusion models generate samples comparable with those produced by standard autoregressive language models.
arXiv Detail & Related papers (2022-11-08T13:30:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.