Previously on the Stories: Recap Snippet Identification for Story
Reading
- URL: http://arxiv.org/abs/2402.07271v1
- Date: Sun, 11 Feb 2024 18:27:14 GMT
- Title: Previously on the Stories: Recap Snippet Identification for Story
Reading
- Authors: Jiangnan Li, Qiujing Wang, Liyan Xu, Wenjie Pang, Mo Yu, Zheng Lin,
Weiping Wang, Jie Zhou
- Abstract summary: We propose the first benchmark on this useful task called Recap Snippet Identification with a hand-crafted evaluation dataset.
Our experiments show that the proposed task is challenging to PLMs, LLMs, and proposed methods as the task requires a deep understanding of the plot correlation between snippets.
- Score: 51.641565531840186
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Similar to the "previously-on" scenes in TV shows, recaps can help book
reading by recalling the readers' memory about the important elements in
previous texts to better understand the ongoing plot. Despite its usefulness,
this application has not been well studied in the NLP community. We propose the
first benchmark on this useful task called Recap Snippet Identification with a
hand-crafted evaluation dataset. Our experiments show that the proposed task is
challenging to PLMs, LLMs, and proposed methods as the task requires a deep
understanding of the plot correlation between snippets.
Related papers
- Generative Retrieval with Large Language Models [8.069852420775362]
This paper explores leveraging the parameterized knowledge stored during the pre-training phase of large language models to independently recall reference passage from any starting position.
Experiments on KILT knowledge-sensitive tasks have verified that LLMs can independently recall reference passage location in various task forms.
arXiv Detail & Related papers (2024-02-26T20:35:32Z) - Benchmarking LLMs on the Semantic Overlap Summarization Task [9.656095701778975]
This paper comprehensively evaluates Large Language Models (LLMs) on the Semantic Overlap Summarization (SOS) task.
We report well-established metrics like ROUGE, BERTscore, and SEM-F1$ on two different datasets of alternative narratives.
arXiv Detail & Related papers (2024-02-26T20:33:50Z) - Walking Down the Memory Maze: Beyond Context Limit through Interactive
Reading [63.93888816206071]
We introduce MemWalker, a method that processes the long context into a tree of summary nodes. Upon receiving a query, the model navigates this tree in search of relevant information, and responds once it gathers sufficient information.
We show that, beyond effective reading, MemWalker enhances explainability by highlighting the reasoning steps as it interactively reads the text; pinpointing the relevant text segments related to the query.
arXiv Detail & Related papers (2023-10-08T06:18:14Z) - Query Rewriting for Retrieval-Augmented Large Language Models [139.242907155883]
Large Language Models (LLMs) play powerful, black-box readers in the retrieve-then-read pipeline.
This work introduces a new framework, Rewrite-Retrieve-Read instead of the previous retrieve-then-read for the retrieval-augmented LLMs.
arXiv Detail & Related papers (2023-05-23T17:27:50Z) - Improving Keyphrase Extraction with Data Augmentation and Information
Filtering [67.43025048639333]
Keyphrase extraction is one of the essential tasks for document understanding in NLP.
We present a novel corpus and method for keyphrase extraction from the videos streamed on the Behance platform.
arXiv Detail & Related papers (2022-09-11T22:38:02Z) - Learning to Ask Conversational Questions by Optimizing Levenshtein
Distance [83.53855889592734]
We introduce a Reinforcement Iterative Sequence Editing (RISE) framework that optimize the minimum Levenshtein distance (MLD) through explicit editing actions.
RISE is able to pay attention to tokens that are related to conversational characteristics.
Experimental results on two benchmark datasets show that RISE significantly outperforms state-of-the-art methods.
arXiv Detail & Related papers (2021-06-30T08:44:19Z) - ReadTwice: Reading Very Large Documents with Memories [19.45538971299312]
We propose ReadTwice, a technique that combines several strengths of prior approaches to model long-range dependencies with Transformers.
The main idea is to read text in small segments, in parallel, summarizing each segment into a memory table to be used in a second read of the text.
We show that the method outperforms models of comparable size on several question answering (QA) datasets and sets a new state of the art on the challenging NarrativeQA task.
arXiv Detail & Related papers (2021-05-10T10:13:09Z) - Hone as You Read: A Practical Type of Interactive Summarization [6.662800021628275]
We present HARE, a new task where reader feedback is used to optimize document summaries for personal interest.
This task is related to interactive summarization, where personalized summaries are produced following a long feedback stage.
We propose to gather minimally-invasive feedback during the reading process to adapt to user interests and augment the document in real-time.
arXiv Detail & Related papers (2021-05-06T19:36:40Z) - Exploring Content Selection in Summarization of Novel Chapters [19.11830806780343]
We present a new summarization task, generating summaries of novel chapters using summary/chapter pairs from online study guides.
This is a harder task than the news summarization task, given the chapter length as well as the extreme paraphrasing and generalization found in the summaries.
We focus on extractive summarization, which requires the creation of a gold-standard set of extractive summaries.
arXiv Detail & Related papers (2020-05-04T20:45:39Z) - The Shmoop Corpus: A Dataset of Stories with Loosely Aligned Summaries [72.48439126769627]
We introduce the Shmoop Corpus: a dataset of 231 stories paired with detailed multi-paragraph summaries for each individual chapter.
From the corpus, we construct a set of common NLP tasks, including Cloze-form question answering and a simplified form of abstractive summarization.
We believe that the unique structure of this corpus provides an important foothold towards making machine story comprehension more approachable.
arXiv Detail & Related papers (2019-12-30T21:03:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.