Generaci\'on autom\'atica de frases literarias en espa\~nol
- URL: http://arxiv.org/abs/2001.11381v1
- Date: Fri, 17 Jan 2020 15:42:14 GMT
- Title: Generaci\'on autom\'atica de frases literarias en espa\~nol
- Authors: Luis-Gil Moreno-Jim\'enez, Juan-Manuel Torres-Moreno, Roseli S.
Wedemann
- Abstract summary: We address the automatic generation of literary sentences in Spanish.
We propose three models of text generation based mainly on statistical algorithms and shallow parsing analysis.
- Score: 1.2998637003026272
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we present a state of the art in the area of Computational
Creativity (CC). In particular, we address the automatic generation of literary
sentences in Spanish. We propose three models of text generation based mainly
on statistical algorithms and shallow parsing analysis. We also present some
rather encouraging preliminary results.
Related papers
- GPT Czech Poet: Generation of Czech Poetic Strophes with Language Models [0.4444634303550442]
We introduce a new model for generating poetry in Czech language, based on fine-tuning a pre-trained Large Language Model.
We demonstrate that guiding the generation process by explicitly specifying strophe parameters within the poem text strongly improves the effectiveness of the model.
arXiv Detail & Related papers (2024-06-18T06:19:45Z) - Spotting AI's Touch: Identifying LLM-Paraphrased Spans in Text [61.22649031769564]
We propose a novel framework, paraphrased text span detection (PTD)
PTD aims to identify paraphrased text spans within a text.
We construct a dedicated dataset, PASTED, for paraphrased text span detection.
arXiv Detail & Related papers (2024-05-21T11:22:27Z) - A User-Centered Evaluation of Spanish Text Simplification [6.046875672600245]
We present an evaluation of text simplification (TS) in Spanish for a production system.
We compare the most prevalent Spanish-specific readability scores with neural networks, and show that the latter are consistently better at predicting user preferences regarding TS.
We release the corpora in our evaluation to the broader community with the hopes of pushing forward the state-of-the-art in Spanish natural language processing.
arXiv Detail & Related papers (2023-08-15T03:49:59Z) - Disco-Bench: A Discourse-Aware Evaluation Benchmark for Language
Modelling [70.23876429382969]
We propose a benchmark that can evaluate intra-sentence discourse properties across a diverse set of NLP tasks.
Disco-Bench consists of 9 document-level testsets in the literature domain, which contain rich discourse phenomena.
For linguistic analysis, we also design a diagnostic test suite that can examine whether the target models learn discourse knowledge.
arXiv Detail & Related papers (2023-07-16T15:18:25Z) - A Survey on Retrieval-Augmented Text Generation [53.04991859796971]
Retrieval-augmented text generation has remarkable advantages and has achieved state-of-the-art performance in many NLP tasks.
It firstly highlights the generic paradigm of retrieval-augmented generation, and then it reviews notable approaches according to different tasks.
arXiv Detail & Related papers (2022-02-02T16:18:41Z) - A Preliminary Study for Literary Rhyme Generation based on Neuronal
Representation, Semantics and Shallow Parsing [1.7188280334580195]
We introduce a model for the generation of literary rhymes in Spanish, combining structures of language and neural network models.
Results obtained with a manual evaluation of the texts generated by our algorithm are encouraging.
arXiv Detail & Related papers (2021-12-25T14:40:09Z) - How much do language models copy from their training data? Evaluating
linguistic novelty in text generation using RAVEN [63.79300884115027]
Current language models can generate high-quality text.
Are they simply copying text they have seen before, or have they learned generalizable linguistic abstractions?
We introduce RAVEN, a suite of analyses for assessing the novelty of generated text.
arXiv Detail & Related papers (2021-11-18T04:07:09Z) - Quasi Error-free Text Classification and Authorship Recognition in a
large Corpus of English Literature based on a Novel Feature Set [0.0]
We show that in the entire GLEC quasi error-free text classification and authorship recognition is possible with a method using the same set of five style and five content features.
Our data pave the way for many future computational and empirical studies of literature or experiments in reading psychology.
arXiv Detail & Related papers (2020-10-21T07:39:55Z) - Unsupervised Text Generation by Learning from Search [86.51619839836331]
TGLS is a novel framework to unsupervised Text Generation by Learning.
We demonstrate the effectiveness of TGLS on two real-world natural language generation tasks, paraphrase generation and text formalization.
arXiv Detail & Related papers (2020-07-09T04:34:48Z) - Enabling Language Models to Fill in the Blanks [81.59381915581892]
We present a simple approach for text infilling, the task of predicting missing spans of text at any position in a document.
We train (or fine-tune) off-the-shelf language models on sequences containing the concatenation of artificially-masked text and the text which was masked.
We show that this approach, which we call infilling by language modeling, can enable LMs to infill entire sentences effectively on three different domains: short stories, scientific abstracts, and lyrics.
arXiv Detail & Related papers (2020-05-11T18:00:03Z) - QURIOUS: Question Generation Pretraining for Text Generation [13.595014409069584]
We propose question generation as a pretraining method, which better aligns with the text generation objectives.
Our text generation models pretrained with this method are better at understanding the essence of the input and are better language models for the target task.
arXiv Detail & Related papers (2020-04-23T08:41:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.