Uniform Complexity for Text Generation
- URL: http://arxiv.org/abs/2204.05185v3
- Date: Thu, 19 Oct 2023 21:40:18 GMT
- Title: Uniform Complexity for Text Generation
- Authors: Joseph Marvin Imperial, Harish Tayyar Madabushi
- Abstract summary: We introduce Uniform Complexity for Text Generation (UCTG), a new benchmark test which raises the challenge of making generative models observe uniform linguistic properties with respect to prompts.
We find that models such as GPT-2 struggle to preserve the complexity of input prompts used in its generations, even if finetuned with professionally written texts.
- Score: 4.867923281108005
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large language models (LLMs) have shown promising results in a wide array of
generative NLP tasks, such as summarization and machine translation. In the
context of narrative generation, however, existing models still do not capture
factors that contribute to producing consistent text. For instance, it is
logical that a piece of text or a story should be uniformly readable throughout
and that this form of complexity should be controllable. As such, if the
complexity of an input text prompt is rated first-grade reading level in the
Flesch Reading Ease test, then the generated text continuing the plot should
also be within this range of complexity. With this in mind, we introduce
Uniform Complexity for Text Generation (UCTG), a new benchmark test which
raises the challenge of making generative models observe uniform linguistic
properties with respect to prompts. We experiment with over 150+ linguistically
and cognitively motivated features for evaluating text complexity in humans and
generative models. From our results, we find that models such as GPT-2 struggle
to preserve the complexity of input prompts used in its generations, even if
finetuned with professionally written texts.
Related papers
- DPIC: Decoupling Prompt and Intrinsic Characteristics for LLM Generated Text Detection [56.513637720967566]
Large language models (LLMs) can generate texts that pose risks of misuse, such as plagiarism, planting fake reviews on e-commerce platforms, or creating inflammatory false tweets.
Existing high-quality detection methods usually require access to the interior of the model to extract the intrinsic characteristics.
We propose to extract deep intrinsic characteristics of the black-box model generated texts.
arXiv Detail & Related papers (2023-05-21T17:26:16Z) - MURMUR: Modular Multi-Step Reasoning for Semi-Structured Data-to-Text
Generation [102.20036684996248]
We propose MURMUR, a neuro-symbolic modular approach to text generation from semi-structured data with multi-step reasoning.
We conduct experiments on two data-to-text generation tasks like WebNLG and LogicNLG.
arXiv Detail & Related papers (2022-12-16T17:36:23Z) - Lexical Complexity Controlled Sentence Generation [6.298911438929862]
We introduce a novel task of lexical complexity controlled sentence generation.
It has enormous potential in domains such as grade reading, language teaching and acquisition.
We propose a simple but effective approach for this task based on complexity embedding.
arXiv Detail & Related papers (2022-11-26T11:03:56Z) - A Benchmark Corpus for the Detection of Automatically Generated Text in
Academic Publications [0.02578242050187029]
This paper presents two datasets comprised of artificially generated research content.
In the first case, the content is completely generated by the GPT-2 model after a short prompt extracted from original papers.
The partial or hybrid dataset is created by replacing several sentences of abstracts with sentences that are generated by the Arxiv-NLP model.
We evaluate the quality of the datasets comparing the generated texts to aligned original texts using fluency metrics such as BLEU and ROUGE.
arXiv Detail & Related papers (2022-02-04T08:16:56Z) - How much do language models copy from their training data? Evaluating
linguistic novelty in text generation using RAVEN [63.79300884115027]
Current language models can generate high-quality text.
Are they simply copying text they have seen before, or have they learned generalizable linguistic abstractions?
We introduce RAVEN, a suite of analyses for assessing the novelty of generated text.
arXiv Detail & Related papers (2021-11-18T04:07:09Z) - Long Text Generation by Modeling Sentence-Level and Discourse-Level
Coherence [59.51720326054546]
We propose a long text generation model, which can represent the prefix sentences at sentence level and discourse level in the decoding process.
Our model can generate more coherent texts than state-of-the-art baselines.
arXiv Detail & Related papers (2021-05-19T07:29:08Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.