Vietnamese Poem Generation & The Prospect Of Cross-Language Poem-To-Poem
Translation
- URL: http://arxiv.org/abs/2401.01078v3
- Date: Thu, 4 Jan 2024 17:29:47 GMT
- Title: Vietnamese Poem Generation & The Prospect Of Cross-Language Poem-To-Poem
Translation
- Authors: Triet Minh Huynh and Quan Le Bao
- Abstract summary: We propose using Large Language Models to generate Vietnamese poems from natural language prompts.
The GPT-3 Babbage variant achieves a custom evaluation score of 0.8, specifically tailored to the "luc bat" genre of Vietnamese poetry.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Poetry generation has been a challenging task in the field of Natural
Language Processing, as it requires the model to understand the nuances of
language, sentiment, and style. In this paper, we propose using Large Language
Models to generate Vietnamese poems of various genres from natural language
prompts, thereby facilitating an intuitive process with enhanced content
control. Our most efficacious model, the GPT-3 Babbage variant, achieves a
custom evaluation score of 0.8, specifically tailored to the "luc bat" genre of
Vietnamese poetry. Furthermore, we also explore the idea of paraphrasing poems
into normal text prompts and yield a relatively high score of 0.781 in the "luc
bat" genre. This experiment presents the potential for cross-Language
poem-to-poem translation with translated poems as the inputs while concurrently
maintaining complete control over the generated content.
Related papers
- PoetryDiffusion: Towards Joint Semantic and Metrical Manipulation in
Poetry Generation [58.36105306993046]
Controllable text generation is a challenging and meaningful field in natural language generation (NLG)
In this paper, we pioneer the use of the Diffusion model for generating sonnets and Chinese SongCi poetry.
Our model outperforms existing models in automatic evaluation of semantic, metrical, and overall performance as well as human evaluation.
arXiv Detail & Related papers (2023-06-14T11:57:31Z) - A Corpus for Sentence-level Subjectivity Detection on English News Articles [49.49218203204942]
We use our guidelines to collect NewsSD-ENG, a corpus of 638 objective and 411 subjective sentences extracted from English news articles on controversial topics.
Our corpus paves the way for subjectivity detection in English without relying on language-specific tools, such as lexicons or machine translation.
arXiv Detail & Related papers (2023-05-29T11:54:50Z) - Translate to Disambiguate: Zero-shot Multilingual Word Sense
Disambiguation with Pretrained Language Models [67.19567060894563]
Pretrained Language Models (PLMs) learn rich cross-lingual knowledge and can be finetuned to perform well on diverse tasks.
We present a new study investigating how well PLMs capture cross-lingual word sense with Contextual Word-Level Translation (C-WLT)
We find that as the model size increases, PLMs encode more cross-lingual word sense knowledge and better use context to improve WLT performance.
arXiv Detail & Related papers (2023-04-26T19:55:52Z) - SP-GPT2: Semantics Improvement in Vietnamese Poetry Generation [1.9107347888374506]
Generative Pretraining Transformer 2 (GPT2) is one of the state of the art approaches that have excellent successes.
In this paper, we took the first step to investigate the power of GPT2 in traditional Vietnamese poetry generation.
We released the first computational scoring module for poems generated in the template containing the style rule dictionary.
arXiv Detail & Related papers (2021-10-10T14:31:08Z) - Prose2Poem: The Blessing of Transformers in Translating Prose to Persian
Poetry [2.15242029196761]
We introduce a novel Neural Machine Translation (NMT) approach to translate prose to ancient Persian poetry.
We trained a Transformer model from scratch to obtain initial translations and pretrained different variations of BERT to obtain final translations.
arXiv Detail & Related papers (2021-09-30T09:04:11Z) - Don't Go Far Off: An Empirical Study on Neural Poetry Translation [13.194404923699782]
We present an empirical investigation for poetry translation along several dimensions.
We contribute a parallel dataset of poetry translations for several language pairs.
Our results show that multilingual fine-tuning on poetic text significantly outperforms multilingual fine-tuning on non-poetic text that is 35X larger in size.
arXiv Detail & Related papers (2021-09-07T10:00:44Z) - CCPM: A Chinese Classical Poetry Matching Dataset [50.90794811956129]
We propose a novel task to assess a model's semantic understanding of poetry by poem matching.
This task requires the model to select one line of Chinese classical poetry among four candidates according to the modern Chinese translation of a line of poetry.
To construct this dataset, we first obtain a set of parallel data of Chinese classical poetry and modern Chinese translation.
arXiv Detail & Related papers (2021-06-03T16:49:03Z) - Generating Major Types of Chinese Classical Poetry in a Uniformed
Framework [88.57587722069239]
We propose a GPT-2 based framework for generating major types of Chinese classical poems.
Preliminary results show this enhanced model can generate Chinese classical poems of major types with high quality in both form and content.
arXiv Detail & Related papers (2020-03-13T14:16:25Z) - MixPoet: Diverse Poetry Generation via Learning Controllable Mixed
Latent Space [79.70053419040902]
We propose MixPoet, a novel model that absorbs multiple factors to create various styles and promote diversity.
Based on a semi-supervised variational autoencoder, our model disentangles the latent space into some subspaces, with each conditioned on one influence factor by adversarial training.
Experiment results on Chinese poetry demonstrate that MixPoet improves both diversity and quality against three state-of-the-art models.
arXiv Detail & Related papers (2020-03-13T03:31:29Z) - Introducing Aspects of Creativity in Automatic Poetry Generation [2.792030485253753]
Poetry Generation involves teaching systems to automatically generate text that resembles poetic work.
A deep learning system can learn to generate poetry on its own by training on a corpus of poems and modeling the particular style of language.
We propose taking an approach that fine-tunes GPT-2, a pre-trained language model, to our downstream task of poetry generation.
arXiv Detail & Related papers (2020-02-06T20:44:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.