Event Segmentation Applications in Large Language Model Enabled Automated Recall Assessments
- URL: http://arxiv.org/abs/2502.13349v1
- Date: Wed, 19 Feb 2025 00:48:51 GMT
- Title: Event Segmentation Applications in Large Language Model Enabled Automated Recall Assessments
- Authors: Ryan A. Panela, Alex J. Barnett, Morgan D. Barense, Björn Herrmann,
- Abstract summary: Event segmentation is central to how we perceive, encode, and recall experiences.
Current research methodologies rely heavily on human for assessing segmentation patterns and recall ability.
We leverage Large Language Models (LLMs) to automate event segmentation and assess recall.
- Score: 0.0
- License:
- Abstract: Understanding how individuals perceive and recall information in their natural environments is critical to understanding potential failures in perception (e.g., sensory loss) and memory (e.g., dementia). Event segmentation, the process of identifying distinct events within dynamic environments, is central to how we perceive, encode, and recall experiences. This cognitive process not only influences moment-to-moment comprehension but also shapes event specific memory. Despite the importance of event segmentation and event memory, current research methodologies rely heavily on human judgements for assessing segmentation patterns and recall ability, which are subjective and time-consuming. A few approaches have been introduced to automate event segmentation and recall scoring, but validity with human responses and ease of implementation require further advancements. To address these concerns, we leverage Large Language Models (LLMs) to automate event segmentation and assess recall, employing chat completion and text-embedding models, respectively. We validated these models against human annotations and determined that LLMs can accurately identify event boundaries, and that human event segmentation is more consistent with LLMs than among humans themselves. Using this framework, we advanced an automated approach for recall assessments which revealed semantic similarity between segmented narrative events and participant recall can estimate recall performance. Our findings demonstrate that LLMs can effectively simulate human segmentation patterns and provide recall evaluations that are a scalable alternative to manual scoring. This research opens novel avenues for studying the intersection between perception, memory, and cognitive impairment using methodologies driven by artificial intelligence.
Related papers
- Episodic Memories Generation and Evaluation Benchmark for Large Language Models [7.660368798066376]
We argue that integrating episodic memory capabilities into Large Language Models is essential for advancing AI towards human-like cognition.
We develop a structured approach to represent episodic events, encapsulating temporal and spatial contexts, involved entities, and detailed descriptions.
We synthesize a unique episodic memory benchmark, free from contamination, and release open source code and datasets to assess LLM performance.
arXiv Detail & Related papers (2025-01-21T02:16:13Z) - Detecting Memorization in Large Language Models [0.0]
Large language models (LLMs) have achieved impressive results in natural language processing but are prone to memorizing portions of their training data.
Traditional methods for detecting memorization rely on output probabilities or loss functions.
We introduce an analytical method that precisely detects memorization by examining neuron activations within the LLM.
arXiv Detail & Related papers (2024-12-02T00:17:43Z) - MADial-Bench: Towards Real-world Evaluation of Memory-Augmented Dialogue Generation [15.64077949677469]
We present a novel Memory-Augmented Dialogue Benchmark (MADail-Bench) to evaluate the effectiveness of memory-augmented dialogue systems (MADS)
The benchmark assesses two tasks separately: memory retrieval and memory recognition with the incorporation of both passive and proactive memory recall data.
Results from cutting-edge embedding models and large language models on this benchmark indicate the potential for further advancement.
arXiv Detail & Related papers (2024-09-23T17:38:41Z) - Modeling User Preferences via Brain-Computer Interfacing [54.3727087164445]
We use Brain-Computer Interfacing technology to infer users' preferences, their attentional correlates towards visual content, and their associations with affective experience.
We link these to relevant applications, such as information retrieval, personalized steering of generative models, and crowdsourcing population estimates of affective experiences.
arXiv Detail & Related papers (2024-05-15T20:41:46Z) - In-Memory Learning: A Declarative Learning Framework for Large Language
Models [56.62616975119192]
We propose a novel learning framework that allows agents to align with their environment without relying on human-labeled data.
This entire process transpires within the memory components and is implemented through natural language.
We demonstrate the effectiveness of our framework and provide insights into this problem.
arXiv Detail & Related papers (2024-03-05T08:25:11Z) - Saliency-Guided Hidden Associative Replay for Continual Learning [13.551181595881326]
Continual Learning is a burgeoning domain in next-generation AI, focusing on training neural networks over a sequence of tasks akin to human learning.
This paper presents the Saliency Guided Hidden Associative Replay for Continual Learning.
This novel framework synergizes associative memory with replay-based strategies. SHARC primarily archives salient data segments via sparse memory encoding.
arXiv Detail & Related papers (2023-10-06T15:54:12Z) - Are All Steps Equally Important? Benchmarking Essentiality Detection of
Events [92.92425231146433]
This paper examines the extent to which current models comprehend the essentiality of step events in relation to a goal event.
We contribute a high-quality corpus of (goal, step) pairs gathered from the community guideline website WikiHow.
The high inter-annotator agreement demonstrates that humans possess a consistent understanding of event essentiality.
arXiv Detail & Related papers (2022-10-08T18:00:22Z) - Predictive Event Segmentation and Representation with Neural Networks: A
Self-Supervised Model Assessed by Psychological Experiments [2.223733768286313]
We introduce a self-supervised model of event segmentation.
Our model consists of neural networks that predict the sensory signal in the next time-step to represent different events.
We show that our model that tracks the prediction error signals can produce human-like event boundaries and event representations.
arXiv Detail & Related papers (2022-10-04T14:14:30Z) - Pin the Memory: Learning to Generalize Semantic Segmentation [68.367763672095]
We present a novel memory-guided domain generalization method for semantic segmentation based on meta-learning framework.
Our method abstracts the conceptual knowledge of semantic classes into categorical memory which is constant beyond the domains.
arXiv Detail & Related papers (2022-04-07T17:34:01Z) - CoSeg: Cognitively Inspired Unsupervised Generic Event Segmentation [118.18977078626776]
We propose an end-to-end self-supervised learning framework for event segmentation/boundary detection.
Our framework exploits a transformer-based feature reconstruction scheme to detect event boundary by reconstruction errors.
The goal of our work is to segment generic events rather than localize some specific ones.
arXiv Detail & Related papers (2021-09-30T14:40:32Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.