Modelling Real-Life Cycling Decisions in Real Urban Settings Through Psychophysiology and LLM-Derived Contextual Data
- URL: http://arxiv.org/abs/2510.24726v1
- Date: Fri, 03 Oct 2025 01:19:45 GMT
- Title: Modelling Real-Life Cycling Decisions in Real Urban Settings Through Psychophysiology and LLM-Derived Contextual Data
- Authors: Maximiliano Rosadio Z., Angel Jimenez-Molina, Bastián Henríquez, Paulina Leiva, Ricardo Hurtubia, Ricardo De La Paz Guala, Leandro Gayozo, C. Angelo Guevara,
- Abstract summary: This paper applies an innovative approach, extracting contextual data from recorded multimedia using large language models (LLMs)<n>The applied models focus on understanding how different environments and traffic situations affect the emotional states and behaviors of the participants using physiological data.<n>The study confirms that cycling decisions are influenced by stress-related emotions and highlights the strong impact of urban characteristics and traffic conditions on cyclist behavior.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Measuring emotional states in transportation contexts is an emerging field. Methods based on self-reported emotions are limited by their low granularity and their susceptibility to memory bias. In contrast, methods based on physiological indicators provide continuous data, enabling researchers to measure changes in emotional states with high detail and accuracy. Not only are emotions important in the analysis, but understanding what triggers emotional changes is equally important. Uncontrolled variables such as traffic conditions, pedestrian interactions, and infrastructure remain a significant challenge, as they can have a great impact on emotional states. Explaining the reasons behind these emotional states requires gathering sufficient and proper contextual data, which can be extremely difficult in real-world environments. This paper addresses these challenges by applying an innovative approach, extracting contextual data (expert annotator level) from recorded multimedia using large language models (LLMs). In this paper, data are collected from an urban cycling case study of the City of Santiago, Chile. The applied models focus on understanding how different environments and traffic situations affect the emotional states and behaviors of the participants using physiological data. Sequences of images, extracted from the recorded videos, are processed by LLMs to obtain semantic descriptions of the environment. These discrete, although dense and detailed, contextual data are integrated into a hybrid model, where fatigue and arousal serve as latent variables influencing observed cycling behaviors (inferred from GPS data) like waiting, accelerating, braking, etc. The study confirms that cycling decisions are influenced by stress-related emotions and highlights the strong impact of urban characteristics and traffic conditions on cyclist behavior.
Related papers
- How built environment shapes cycling experience: A multi-scale review in historical urban contexts [2.770226625653906]
We systematically reviewed 68 studies across urban planning, transportation, behavioural science, neuroscience, and public health.<n>We find a persistent reliance on objective proxies, limited integration of subjective accounts, and insufficient attention to the streetscape as a lived environment.
arXiv Detail & Related papers (2025-09-19T06:30:32Z) - EmoCAST: Emotional Talking Portrait via Emotive Text Description [56.42674612728354]
EmoCAST is a diffusion-based framework for precise text-driven emotional synthesis.<n>In appearance modeling, emotional prompts are integrated through a text-guided decoupled emotive module.<n>EmoCAST achieves state-of-the-art performance in generating realistic, emotionally expressive, and audio-synchronized talking-head videos.
arXiv Detail & Related papers (2025-08-28T10:02:06Z) - CAST-Phys: Contactless Affective States Through Physiological signals Database [74.28082880875368]
The lack of affective multi-modal datasets remains a major bottleneck in developing accurate emotion recognition systems.<n>We present the Contactless Affective States Through Physiological Signals Database (CAST-Phys), a novel high-quality dataset capable of remote physiological emotion recognition.<n>Our analysis highlights the crucial role of physiological signals in realistic scenarios where facial expressions alone may not provide sufficient emotional information.
arXiv Detail & Related papers (2025-07-08T15:20:24Z) - Modelling Emotions in Face-to-Face Setting: The Interplay of Eye-Tracking, Personality, and Temporal Dynamics [1.4645774851707578]
In this study, we showcase how integrating eye-tracking data, temporal dynamics, and personality traits can substantially enhance the detection of both perceived and felt emotions.<n>Our findings inform the design of future affective computing and human-agent systems.
arXiv Detail & Related papers (2025-03-18T13:15:32Z) - Implicit Design Choices and Their Impact on Emotion Recognition Model
Development and Evaluation [5.534160116442057]
The subjectivity of emotions poses significant challenges in developing accurate and robust computational models.
This thesis examines critical facets of emotion recognition, beginning with the collection of diverse datasets.
To handle the challenge of non-representative training data, this work collects the Multimodal Stressed Emotion dataset.
arXiv Detail & Related papers (2023-09-06T02:45:42Z) - I am Only Happy When There is Light: The Impact of Environmental Changes
on Affective Facial Expressions Recognition [65.69256728493015]
We study the impact of different image conditions on the recognition of arousal from human facial expressions.
Our results show how the interpretation of human affective states can differ greatly in either the positive or negative direction.
arXiv Detail & Related papers (2022-10-28T16:28:26Z) - Exploring Contextual Representation and Multi-Modality for End-to-End
Autonomous Driving [58.879758550901364]
Recent perception systems enhance spatial understanding with sensor fusion but often lack full environmental context.
We introduce a framework that integrates three cameras to emulate the human field of view, coupled with top-down bird-eye-view semantic data to enhance contextual representation.
Our method achieves displacement error by 0.67m in open-loop settings, surpassing current methods by 6.9% on the nuScenes dataset.
arXiv Detail & Related papers (2022-10-13T05:56:20Z) - A cross-corpus study on speech emotion recognition [29.582678406878568]
This study investigates whether information learnt from acted emotions is useful for detecting natural emotions.
Four adult English datasets covering acted, elicited and natural emotions are considered.
A state-of-the-art model is proposed to accurately investigate the degradation of performance.
arXiv Detail & Related papers (2022-07-05T15:15:22Z) - The MuSe 2022 Multimodal Sentiment Analysis Challenge: Humor, Emotional
Reactions, and Stress [71.06453250061489]
The Multimodal Sentiment Analysis Challenge (MuSe) 2022 is dedicated to multimodal sentiment and emotion recognition.
For this year's challenge, we feature three datasets: (i) the Passau Spontaneous Football Coach Humor dataset that contains audio-visual recordings of German football coaches, labelled for the presence of humour; (ii) the Hume-Reaction dataset in which reactions of individuals to emotional stimuli have been annotated with respect to seven emotional expression intensities; and (iii) the Ulm-Trier Social Stress Test dataset comprising of audio-visual data labelled with continuous emotion values of people in stressful dispositions.
arXiv Detail & Related papers (2022-06-23T13:34:33Z) - Affect2MM: Affective Analysis of Multimedia Content Using Emotion
Causality [84.69595956853908]
We present Affect2MM, a learning method for time-series emotion prediction for multimedia content.
Our goal is to automatically capture the varying emotions depicted by characters in real-life human-centric situations and behaviors.
arXiv Detail & Related papers (2021-03-11T09:07:25Z) - A Spontaneous Driver Emotion Facial Expression (DEFE) Dataset for
Intelligent Vehicles [13.877751298110148]
This paper introduces a new dataset, the driver emotion facial expression (DEFE) dataset, for driver spontaneous emotions analysis.
The dataset includes facial expression recordings from 60 participants during driving.
To the best of our knowledge, this is currently the only public driver facial expression dataset.
arXiv Detail & Related papers (2020-04-26T07:15:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.