Tell, Don't Show: Leveraging Language Models' Abstractive Retellings to Model Literary Themes
- URL: http://arxiv.org/abs/2505.23166v1
- Date: Thu, 29 May 2025 06:59:21 GMT
- Title: Tell, Don't Show: Leveraging Language Models' Abstractive Retellings to Model Literary Themes
- Authors: Li Lucy, Camilla Griffiths, Sarah Levine, Jennifer L. Eberhardt, Dorottya Demszky, David Bamman,
- Abstract summary: We propose Retell, a simple, accessible topic modeling approach for literature.<n>We prompt resource-efficient, generative language models (LMs) to tell what passages show.
- Score: 9.471374217162843
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conventional bag-of-words approaches for topic modeling, like latent Dirichlet allocation (LDA), struggle with literary text. Literature challenges lexical methods because narrative language focuses on immersive sensory details instead of abstractive description or exposition: writers are advised to "show, don't tell." We propose Retell, a simple, accessible topic modeling approach for literature. Here, we prompt resource-efficient, generative language models (LMs) to tell what passages show, thereby translating narratives' surface forms into higher-level concepts and themes. By running LDA on LMs' retellings of passages, we can obtain more precise and informative topics than by running LDA alone or by directly asking LMs to list topics. To investigate the potential of our method for cultural analytics, we compare our method's outputs to expert-guided annotations in a case study on racial/cultural identity in high school English language arts books.
Related papers
- Do LLMs Understand Why We Write Diaries? A Method for Purpose Extraction and Clustering [41.94295877935867]
This study introduces a novel method based on Large Language Models (LLMs) to identify and cluster the various purposes of diary writing.<n>Our approach is applied to Soviet-era diaries (1922-1929) from the Prozhito digital archive.
arXiv Detail & Related papers (2025-06-01T12:38:01Z) - Looking for the Inner Music: Probing LLMs' Understanding of Literary Style [3.5757761767474876]
Authorial style is easier to define than genre-level style.<n> pronoun usage and word order prove significant for defining both kinds of literary style.
arXiv Detail & Related papers (2025-02-05T22:20:17Z) - Chain-of-MetaWriting: Linguistic and Textual Analysis of How Small Language Models Write Young Students Texts [0.8192907805418583]
This paper introduces a fine-grained linguistic and textual analysis of multilingual Small Language Models' (SLMs) writing.<n>We mainly focused on short story and essay writing tasks in French for schoolchildren and undergraduate students respectively.
arXiv Detail & Related papers (2024-12-19T15:58:53Z) - Explingo: Explaining AI Predictions using Large Language Models [47.21393184176602]
Large Language Models (LLMs) can transform explanations into human-readable, narrative formats that align with natural communication.<n>The Narrator takes in ML explanations and transforms them into natural-language descriptions.<n>The Grader scores these narratives on a set of metrics including accuracy, completeness, fluency, and conciseness.<n>The findings from this work have been integrated into an open-source tool that makes narrative explanations available for further applications.
arXiv Detail & Related papers (2024-12-06T16:01:30Z) - Narrative Analysis of True Crime Podcasts With Knowledge Graph-Augmented Large Language Models [8.78598447041169]
Large language models (LLMs) still struggle with complex narrative arcs as well as narratives containing conflicting information.
Recent work indicates LLMs augmented with external knowledge bases can improve the accuracy and interpretability of the resulting models.
In this work, we analyze the effectiveness of applying knowledge graphs (KGs) in understanding true-crime podcast data.
arXiv Detail & Related papers (2024-11-01T21:49:00Z) - Are Large Language Models Capable of Generating Human-Level Narratives? [114.34140090869175]
This paper investigates the capability of LLMs in storytelling, focusing on narrative development and plot progression.
We introduce a novel computational framework to analyze narratives through three discourse-level aspects.
We show that explicit integration of discourse features can enhance storytelling, as is demonstrated by over 40% improvement in neural storytelling.
arXiv Detail & Related papers (2024-07-18T08:02:49Z) - LFED: A Literary Fiction Evaluation Dataset for Large Language Models [58.85989777743013]
We collect 95 literary fictions that are either originally written in Chinese or translated into Chinese, covering a wide range of topics across several centuries.
We define a question taxonomy with 8 question categories to guide the creation of 1,304 questions.
We conduct an in-depth analysis to ascertain how specific attributes of literary fictions (e.g., novel types, character numbers, the year of publication) impact LLM performance in evaluations.
arXiv Detail & Related papers (2024-05-16T15:02:24Z) - Large Language Models Offer an Alternative to the Traditional Approach of Topic Modelling [0.9095496510579351]
We investigate the untapped potential of large language models (LLMs) as an alternative for uncovering the underlying topics within extensive text corpora.
Our findings indicate that LLMs with appropriate prompts can stand out as a viable alternative, capable of generating relevant topic titles and adhering to human guidelines to refine and merge topics.
arXiv Detail & Related papers (2024-03-24T17:39:51Z) - Testing the Ability of Language Models to Interpret Figurative Language [69.59943454934799]
Figurative and metaphorical language are commonplace in discourse.
It remains an open question to what extent modern language models can interpret nonliteral phrases.
We introduce Fig-QA, a Winograd-style nonliteral language understanding task.
arXiv Detail & Related papers (2022-04-26T23:42:22Z) - Towards Language Modelling in the Speech Domain Using Sub-word
Linguistic Units [56.52704348773307]
We propose a novel LSTM-based generative speech LM based on linguistic units including syllables and phonemes.
With a limited dataset, orders of magnitude smaller than that required by contemporary generative models, our model closely approximates babbling speech.
We show the effect of training with auxiliary text LMs, multitask learning objectives, and auxiliary articulatory features.
arXiv Detail & Related papers (2021-10-31T22:48:30Z) - Topic Adaptation and Prototype Encoding for Few-Shot Visual Storytelling [81.33107307509718]
We propose a topic adaptive storyteller to model the ability of inter-topic generalization.
We also propose a prototype encoding structure to model the ability of intra-topic derivation.
Experimental results show that topic adaptation and prototype encoding structure mutually bring benefit to the few-shot model.
arXiv Detail & Related papers (2020-08-11T03:55:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.