FaNS: a Facet-based Narrative Similarity Metric
- URL: http://arxiv.org/abs/2309.04823v2
- Date: Sat, 2 Mar 2024 14:46:27 GMT
- Title: FaNS: a Facet-based Narrative Similarity Metric
- Authors: Mousumi Akter, Shubhra Kanti Karmaker Santu
- Abstract summary: This paper proposes a novel narrative similarity metric called Facet-based Narrative Similarity (FaNS)
FaNS is based on the classic 5W1H facets (Who, What, When, Where, Why, and How) which are extracted by leveraging the state-of-the-art Large Language Models (LLMs)
- Score: 6.992767260794627
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Similar Narrative Retrieval is a crucial task since narratives are essential
for explaining and understanding events, and multiple related narratives often
help to create a holistic view of the event of interest. To accurately identify
semantically similar narratives, this paper proposes a novel narrative
similarity metric called Facet-based Narrative Similarity (FaNS), based on the
classic 5W1H facets (Who, What, When, Where, Why, and How), which are extracted
by leveraging the state-of-the-art Large Language Models (LLMs). Unlike
existing similarity metrics that only focus on overall lexical/semantic match,
FaNS provides a more granular matching along six different facets independently
and then combines them. To evaluate FaNS, we created a comprehensive dataset by
collecting narratives from AllSides, a third-party news portal. Experimental
results demonstrate that the FaNS metric exhibits a higher correlation (37\%
higher) than traditional text similarity metrics that directly measure the
lexical/semantic match between narratives, demonstrating its effectiveness in
comparing the finer details between a pair of narratives.
Related papers
- Mapping News Narratives Using LLMs and Narrative-Structured Text Embeddings [0.0]
We introduce a numerical narrative representation grounded in structuralist linguistic theory.
We extract the actants using an open-source LLM and integrate them into a Narrative-Structured Text Embedding.
We demonstrate the analytical insights of the method on the example of 5000 full-text news articles from Al Jazeera and The Washington Post on the Israel-Palestine conflict.
arXiv Detail & Related papers (2024-09-10T14:15:30Z) - Are Large Language Models Capable of Generating Human-Level Narratives? [114.34140090869175]
This paper investigates the capability of LLMs in storytelling, focusing on narrative development and plot progression.
We introduce a novel computational framework to analyze narratives through three discourse-level aspects.
We show that explicit integration of discourse features can enhance storytelling, as is demonstrated by over 40% improvement in neural storytelling.
arXiv Detail & Related papers (2024-07-18T08:02:49Z) - GNAT: A General Narrative Alignment Tool [12.100007440638667]
We develop a general approach to narrative alignment coupling the Smith-Waterman algorithm from bioinformatics with modern text similarity metrics.
We apply and evaluate our general narrative alignment tool (GNAT) on four distinct problem domains differing greatly in both the relative and absolute length of documents.
arXiv Detail & Related papers (2023-11-07T00:24:14Z) - StoryAnalogy: Deriving Story-level Analogies from Large Language Models
to Unlock Analogical Understanding [72.38872974837462]
We evaluate the ability to identify and generate analogies by constructing a first-of-its-kind large-scale story-level analogy corpus.
textscStory Analogy contains 24K story pairs from diverse domains with human annotations on two similarities from the extended Structure-Mapping Theory.
We observe that the data in textscStory Analogy can improve the quality of analogy generation in large language models.
arXiv Detail & Related papers (2023-10-19T16:29:23Z) - Discovering collective narratives shifts in online discussions [3.6231158294409482]
We propose a systematic narrative discovery framework that fills the gap by combining change point detection, semantic role labeling (SRL), and automatic aggregation of narrative fragments into narrative networks.
We evaluate our model with synthetic and empirical data two-Twitter corpora about COVID-19 and 2017 French Election.
Results demonstrate that our approach can recover major narrative shifts that correspond to the major events.
arXiv Detail & Related papers (2023-07-17T15:00:04Z) - RoViST:Learning Robust Metrics for Visual Storytelling [2.7124743347047033]
We propose 3 evaluation metrics sets that analyses which aspects we would look for in a good story.
We measure the reliability of our metric sets by analysing its correlation with human judgement scores on a sample of machine stories.
arXiv Detail & Related papers (2022-05-08T03:51:22Z) - Computational Lens on Cognition: Study Of Autobiographical Versus
Imagined Stories With Large-Scale Language Models [95.88620740809004]
We study differences in the narrative flow of events in autobiographical versus imagined stories using GPT-3.
We found that imagined stories have higher sequentiality than autobiographical stories.
In comparison to imagined stories, autobiographical stories contain more concrete words and words related to the first person.
arXiv Detail & Related papers (2022-01-07T20:10:47Z) - Relation Clustering in Narrative Knowledge Graphs [71.98234178455398]
relational sentences in the original text are embedded (with SBERT) and clustered in order to merge together semantically similar relations.
Preliminary tests show that such clustering might successfully detect similar relations, and provide a valuable preprocessing for semi-supervised approaches.
arXiv Detail & Related papers (2020-11-27T10:43:04Z) - Paragraph-level Commonsense Transformers with Recurrent Memory [77.4133779538797]
We train a discourse-aware model that incorporates paragraph-level information to generate coherent commonsense inferences from narratives.
Our results show that PARA-COMET outperforms the sentence-level baselines, particularly in generating inferences that are both coherent and novel.
arXiv Detail & Related papers (2020-10-04T05:24:12Z) - Exploring aspects of similarity between spoken personal narratives by
disentangling them into narrative clause types [13.350982138577038]
We introduce a corpus of real-world spoken personal narratives comprising 10,296 narrative clauses from 594 video transcripts.
Second, we ask non-narrative experts to annotate those clauses under Labov's sociolinguistic model of personal narratives.
Third, we train a classifier that reaches 84.7% F-score for the highest-agreed clauses.
Our approach is intended to help inform machine learning methods aimed at studying or representing personal narratives.
arXiv Detail & Related papers (2020-05-26T14:34:07Z) - Screenplay Summarization Using Latent Narrative Structure [78.45316339164133]
We propose to explicitly incorporate the underlying structure of narratives into general unsupervised and supervised extractive summarization models.
We formalize narrative structure in terms of key narrative events (turning points) and treat it as latent in order to summarize screenplays.
Experimental results on the CSI corpus of TV screenplays, which we augment with scene-level summarization labels, show that latent turning points correlate with important aspects of a CSI episode.
arXiv Detail & Related papers (2020-04-27T11:54:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.