SP-GPT2: Semantics Improvement in Vietnamese Poetry Generation
- URL: http://arxiv.org/abs/2110.15723v1
- Date: Sun, 10 Oct 2021 14:31:08 GMT
- Title: SP-GPT2: Semantics Improvement in Vietnamese Poetry Generation
- Authors: Tuan Nguyen, Hanh Pham, Truong Bui, Tan Nguyen, Duc Luong, Phong
Nguyen
- Abstract summary: Generative Pretraining Transformer 2 (GPT2) is one of the state of the art approaches that have excellent successes.
In this paper, we took the first step to investigate the power of GPT2 in traditional Vietnamese poetry generation.
We released the first computational scoring module for poems generated in the template containing the style rule dictionary.
- Score: 1.9107347888374506
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Automatic text generation has garnered growing attention in recent years as
an essential step towards computer creativity. Generative Pretraining
Transformer 2 (GPT2) is one of the state of the art approaches that have
excellent successes. In this paper, we took the first step to investigate the
power of GPT2 in traditional Vietnamese poetry generation. In the earlier time,
our experiment with base GPT2 was quite good at generating the poem in the
proper template. Though it can learn the patterns, including rhyme and tone
rules, from the training data, like almost all other text generation
approaches, the poems generated still has a topic drift and semantic
inconsistency. To improve the cohesion within the poems, we proposed a new
model SP-GPT2 (semantic poem GPT2) which was built on the top GPT2 model and an
additional loss to constrain context throughout the entire poem. For better
evaluation, we examined the methods by both automatic quantitative evaluation
and human evaluation. Both automatic and human evaluation demonstrated that our
approach can generate poems that have better cohesion without losing the
quality due to additional loss. At the same time, we are the pioneers of this
topic. We released the first computational scoring module for poems generated
in the template containing the style rule dictionary. Additionally, we are the
first to publish a Luc-Bat dataset, including 87609 Luc Bat poems, which is
equivalent to about 2.6 million sentences, combined with about 83579 poems in
other styles was also published for further exploration. The code is available
at https://github.com/fsoft-ailab/Poem-Generator
Related papers
- Does ChatGPT Have a Poetic Style? [0.6827423171182154]
We prompt the GPT-3.5 and GPT-4 models to generate English-language poems in 24 different poetic forms and styles.
We analyze the resulting 5.7k poems, comparing them to a sample of 3.7k poems from the Poetry Foundation and the Academy of American Poets.
We find that the GPT models, especially GPT-4, can successfully produce poems in a range of both common and uncommon English-language forms.
arXiv Detail & Related papers (2024-10-20T06:01:34Z) - Vietnamese Poem Generation & The Prospect Of Cross-Language Poem-To-Poem
Translation [0.0]
We propose using Large Language Models to generate Vietnamese poems from natural language prompts.
The GPT-3 Babbage variant achieves a custom evaluation score of 0.8, specifically tailored to the "luc bat" genre of Vietnamese poetry.
arXiv Detail & Related papers (2024-01-02T07:46:34Z) - PoetryDiffusion: Towards Joint Semantic and Metrical Manipulation in
Poetry Generation [58.36105306993046]
Controllable text generation is a challenging and meaningful field in natural language generation (NLG)
In this paper, we pioneer the use of the Diffusion model for generating sonnets and Chinese SongCi poetry.
Our model outperforms existing models in automatic evaluation of semantic, metrical, and overall performance as well as human evaluation.
arXiv Detail & Related papers (2023-06-14T11:57:31Z) - Unsupervised Melody-to-Lyric Generation [91.29447272400826]
We propose a method for generating high-quality lyrics without training on any aligned melody-lyric data.
We leverage the segmentation and rhythm alignment between melody and lyrics to compile the given melody into decoding constraints.
Our model can generate high-quality lyrics that are more on-topic, singable, intelligible, and coherent than strong baselines.
arXiv Detail & Related papers (2023-05-30T17:20:25Z) - GPoeT-2: A GPT-2 Based Poem Generator [0.5156484100374058]
GPoeT-2 is based on fine-tuning a state of the art natural language model (i.e. GPT-2) to generate limericks.
We present a collection of 94 categorized limericks that rank highly on the explored "good poetry" metrics to provoke human creativity.
arXiv Detail & Related papers (2022-05-18T10:25:12Z) - CCPM: A Chinese Classical Poetry Matching Dataset [50.90794811956129]
We propose a novel task to assess a model's semantic understanding of poetry by poem matching.
This task requires the model to select one line of Chinese classical poetry among four candidates according to the modern Chinese translation of a line of poetry.
To construct this dataset, we first obtain a set of parallel data of Chinese classical poetry and modern Chinese translation.
arXiv Detail & Related papers (2021-06-03T16:49:03Z) - Acrostic Poem Generation [26.604889384391726]
We propose a new task in the area of computational creativity: acrostic poem generation in English.
Acrostic poems are poems that contain a hidden message; typically, the first letter of each line spells out a word or short phrase.
Our experiments show that the acrostic poems generated by our baseline are received well by humans and do not lose much quality due to the additional constraints.
arXiv Detail & Related papers (2020-10-05T18:00:15Z) - SongNet: Rigid Formats Controlled Text Generation [51.428634666559724]
We propose a simple and elegant framework named SongNet to tackle this problem.
The backbone of the framework is a Transformer-based auto-regressive language model.
A pre-training and fine-tuning framework is designed to further improve the generation quality.
arXiv Detail & Related papers (2020-04-17T01:40:18Z) - Generating Major Types of Chinese Classical Poetry in a Uniformed
Framework [88.57587722069239]
We propose a GPT-2 based framework for generating major types of Chinese classical poems.
Preliminary results show this enhanced model can generate Chinese classical poems of major types with high quality in both form and content.
arXiv Detail & Related papers (2020-03-13T14:16:25Z) - MixPoet: Diverse Poetry Generation via Learning Controllable Mixed
Latent Space [79.70053419040902]
We propose MixPoet, a novel model that absorbs multiple factors to create various styles and promote diversity.
Based on a semi-supervised variational autoencoder, our model disentangles the latent space into some subspaces, with each conditioned on one influence factor by adversarial training.
Experiment results on Chinese poetry demonstrate that MixPoet improves both diversity and quality against three state-of-the-art models.
arXiv Detail & Related papers (2020-03-13T03:31:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.