Generating Major Types of Chinese Classical Poetry in a Uniformed
Framework
- URL: http://arxiv.org/abs/2003.11528v1
- Date: Fri, 13 Mar 2020 14:16:25 GMT
- Title: Generating Major Types of Chinese Classical Poetry in a Uniformed
Framework
- Authors: Jinyi Hu, Maosong Sun
- Abstract summary: We propose a GPT-2 based framework for generating major types of Chinese classical poems.
Preliminary results show this enhanced model can generate Chinese classical poems of major types with high quality in both form and content.
- Score: 88.57587722069239
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Poetry generation is an interesting research topic in the field of text
generation. As one of the most valuable literary and cultural heritages of
China, Chinese classical poetry is very familiar and loved by Chinese people
from generation to generation. It has many particular characteristics in its
language structure, ranging from form, sound to meaning, thus is regarded as an
ideal testing task for text generation. In this paper, we propose a GPT-2 based
uniformed framework for generating major types of Chinese classical poems. We
define a unified format for formulating all types of training samples by
integrating detailed form information, then present a simple form-stressed
weighting method in GPT-2 to strengthen the control to the form of the
generated poems, with special emphasis on those forms with longer body length.
Preliminary experimental results show this enhanced model can generate Chinese
classical poems of major types with high quality in both form and content,
validating the effectiveness of the proposed strategy. The model has been
incorporated into Jiuge, the most influential Chinese classical poetry
generation system developed by Tsinghua University (Guo et al., 2019).
Related papers
- PoetryDiffusion: Towards Joint Semantic and Metrical Manipulation in
Poetry Generation [58.36105306993046]
Controllable text generation is a challenging and meaningful field in natural language generation (NLG)
In this paper, we pioneer the use of the Diffusion model for generating sonnets and Chinese SongCi poetry.
Our model outperforms existing models in automatic evaluation of semantic, metrical, and overall performance as well as human evaluation.
arXiv Detail & Related papers (2023-06-14T11:57:31Z) - GPoeT-2: A GPT-2 Based Poem Generator [0.5156484100374058]
GPoeT-2 is based on fine-tuning a state of the art natural language model (i.e. GPT-2) to generate limericks.
We present a collection of 94 categorized limericks that rank highly on the explored "good poetry" metrics to provoke human creativity.
arXiv Detail & Related papers (2022-05-18T10:25:12Z) - TransCouplet:Transformer based Chinese Couplet Generation [1.084959821967413]
Chinese couplet is a form of poetry composed of complex syntax with ancient Chinese language.
This paper presents a transformer-based sequence-to-sequence couplet generation model.
We also evaluate the Glyph, PinYin and Part-of-Speech tagging on the couplet grammatical rules.
arXiv Detail & Related papers (2021-12-03T04:34:48Z) - How much do language models copy from their training data? Evaluating
linguistic novelty in text generation using RAVEN [63.79300884115027]
Current language models can generate high-quality text.
Are they simply copying text they have seen before, or have they learned generalizable linguistic abstractions?
We introduce RAVEN, a suite of analyses for assessing the novelty of generated text.
arXiv Detail & Related papers (2021-11-18T04:07:09Z) - Chinese Traditional Poetry Generating System Based on Deep Learning [0.0]
This paper proposes an automatic generation method of Chinese traditional poetry based on deep learning technology.
It extracts keywords from each poem and matches them with the previous text to make the poem conform to the theme.
When a user inputs a paragraph of text, the machine obtains the theme and generates poem sentence by sentence.
arXiv Detail & Related papers (2021-10-24T02:43:03Z) - CCPM: A Chinese Classical Poetry Matching Dataset [50.90794811956129]
We propose a novel task to assess a model's semantic understanding of poetry by poem matching.
This task requires the model to select one line of Chinese classical poetry among four candidates according to the modern Chinese translation of a line of poetry.
To construct this dataset, we first obtain a set of parallel data of Chinese classical poetry and modern Chinese translation.
arXiv Detail & Related papers (2021-06-03T16:49:03Z) - Lattice-BERT: Leveraging Multi-Granularity Representations in Chinese
Pre-trained Language Models [62.41139712595334]
We propose a novel pre-training paradigm for Chinese -- Lattice-BERT.
We construct a lattice graph from the characters and words in a sentence and feed all these text units into transformers.
We show that our model can bring an average increase of 1.5% under the 12-layer setting.
arXiv Detail & Related papers (2021-04-15T02:36:49Z) - AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding
and Generation [22.08457469951396]
AnchiBERT is a pre-trained language model based on the architecture of BERT.
We evaluate AnchiBERT on both language understanding and generation tasks, including poem classification.
arXiv Detail & Related papers (2020-09-24T03:41:13Z) - SongNet: Rigid Formats Controlled Text Generation [51.428634666559724]
We propose a simple and elegant framework named SongNet to tackle this problem.
The backbone of the framework is a Transformer-based auto-regressive language model.
A pre-training and fine-tuning framework is designed to further improve the generation quality.
arXiv Detail & Related papers (2020-04-17T01:40:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.