ROSbag-based Multimodal Affective Dataset for Emotional and Cognitive
States
- URL: http://arxiv.org/abs/2006.05102v2
- Date: Tue, 20 Oct 2020 15:47:48 GMT
- Title: ROSbag-based Multimodal Affective Dataset for Emotional and Cognitive
States
- Authors: Wonse Jo, Shyam Sundar Kannan, Go-Eum Cha, Ahreum Lee, and Byung-Cheol
Min
- Abstract summary: This paper introduces a new ROSbag-based multimodal affective dataset for emotional and cognitive states generated using Robot Operating System (ROS)
We utilized images and sounds from the International Affective Pictures System (IAPS) and the International Affective Digitized Sounds (IADS) to stimulate targeted emotions.
The generated affective dataset consists of 1,602 ROSbag files, and size of the dataset is about 787GB.
- Score: 0.9786690381850356
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces a new ROSbag-based multimodal affective dataset for
emotional and cognitive states generated using Robot Operating System (ROS). We
utilized images and sounds from the International Affective Pictures System
(IAPS) and the International Affective Digitized Sounds (IADS) to stimulate
targeted emotions (happiness, sadness, anger, fear, surprise, disgust, and
neutral), and a dual N-back game to stimulate different levels of cognitive
workload. 30 human subjects participated in the user study; their physiological
data was collected using the latest commercial wearable sensors, behavioral
data was collected using hardware devices such as cameras, and subjective
assessments were carried out through questionnaires. All data was stored in
single ROSbag files rather than in conventional Comma-separated values (CSV)
files. This not only ensures synchronization of signals and videos in a data
set, but also allows researchers to easily analyze and verify their algorithms
by connecting directly to this dataset through ROS. The generated affective
dataset consists of 1,602 ROSbag files, and size of the dataset is about 787GB.
The dataset is made publicly available. We expect that our dataset can be great
resource for many researchers in the fields of affective computing, HCI, and
HRI.
Related papers
- DIPSER: A Dataset for In-Person Student Engagement Recognition in the Wild [1.766742562532995]
In this paper, a novel dataset is introduced, designed to assess student attention within in-person classroom settings.
This dataset encompasses RGB camera data, featuring multiple cameras per student to capture both posture and facial expressions.
A comprehensive suite of attention and emotion labels for each student is provided, generated through self-reporting and evaluations by four different experts.
Our dataset uniquely combines facial and environmental camera data, smartwatch metrics, and includes underrepresented ethnicities in similar datasets, all within in-the-wild, in-person settings.
arXiv Detail & Related papers (2025-02-27T15:50:21Z) - CoPeD-Advancing Multi-Robot Collaborative Perception: A Comprehensive Dataset in Real-World Environments [8.177157078744571]
This paper presents a pioneering and comprehensive real-world multi-robot collaborative perception dataset.
It features raw sensor inputs, pose estimation, and optional high-level perception annotation.
We believe this work will unlock the potential research of high-level scene understanding through multi-modal collaborative perception in multi-robot settings.
arXiv Detail & Related papers (2024-05-23T15:59:48Z) - SER_AMPEL: a multi-source dataset for speech emotion recognition of
Italian older adults [58.49386651361823]
SER_AMPEL is a multi-source dataset for speech emotion recognition (SER)
It is collected with the aim of providing a reference for speech emotion recognition in case of Italian older adults.
The evidence of the need for such a dataset emerges from the analysis of the state of the art.
arXiv Detail & Related papers (2023-11-24T13:47:25Z) - Source-Free Collaborative Domain Adaptation via Multi-Perspective
Feature Enrichment for Functional MRI Analysis [55.03872260158717]
Resting-state MRI functional (rs-fMRI) is increasingly employed in multi-site research to aid neurological disorder analysis.
Many methods have been proposed to reduce fMRI heterogeneity between source and target domains.
But acquiring source data is challenging due to concerns and/or data storage burdens in multi-site studies.
We design a source-free collaborative domain adaptation framework for fMRI analysis, where only a pretrained source model and unlabeled target data are accessible.
arXiv Detail & Related papers (2023-08-24T01:30:18Z) - Towards Multi-User Activity Recognition through Facilitated Training
Data and Deep Learning for Human-Robot Collaboration Applications [2.3274633659223545]
This study proposes an alternative way of gathering data regarding multi-user activity, by collecting data related to single users and merging them in post-processing.
It is possible to make use of data collected in this way for pair HRC settings and get similar performances compared to using training data regarding groups of users recorded under the same settings.
arXiv Detail & Related papers (2023-02-11T19:27:07Z) - Co-Located Human-Human Interaction Analysis using Nonverbal Cues: A
Survey [71.43956423427397]
We aim to identify the nonverbal cues and computational methodologies resulting in effective performance.
This survey differs from its counterparts by involving the widest spectrum of social phenomena and interaction settings.
Some major observations are: the most often used nonverbal cue, computational method, interaction environment, and sensing approach are speaking activity, support vector machines, and meetings composed of 3-4 persons equipped with microphones and cameras, respectively.
arXiv Detail & Related papers (2022-07-20T13:37:57Z) - Egocentric Human-Object Interaction Detection Exploiting Synthetic Data [19.220651860718892]
We consider the problem of detecting Egocentric HumanObject Interactions (EHOIs) in industrial contexts.
We propose a pipeline and a tool to generate photo-realistic synthetic First Person Vision (FPV) images automatically labeled for EHOI detection.
arXiv Detail & Related papers (2022-04-14T15:59:15Z) - Stereoscopic Universal Perturbations across Different Architectures and
Datasets [60.021985610201156]
We study the effect of adversarial perturbations of images on deep stereo matching networks for the disparity estimation task.
We present a method to craft a single set of perturbations that, when added to any stereo image pair in a dataset, can fool a stereo network.
Our perturbations can increase D1-error (akin to fooling rate) of state-of-the-art stereo networks from 1% to as much as 87%.
arXiv Detail & Related papers (2021-12-12T02:11:31Z) - Finding Facial Forgery Artifacts with Parts-Based Detectors [73.08584805913813]
We design a series of forgery detection systems that each focus on one individual part of the face.
We use these detectors to perform detailed empirical analysis on the FaceForensics++, Celeb-DF, and Facebook Deepfake Detection Challenge datasets.
arXiv Detail & Related papers (2021-09-21T16:18:45Z) - Towards Unbiased Visual Emotion Recognition via Causal Intervention [63.74095927462]
We propose a novel Emotion Recognition Network (IERN) to alleviate the negative effects brought by the dataset bias.
A series of designed tests validate the effectiveness of IERN, and experiments on three emotion benchmarks demonstrate that IERN outperforms other state-of-the-art approaches.
arXiv Detail & Related papers (2021-07-26T10:40:59Z) - Analyzing the Influence of Dataset Composition for Emotion Recognition [0.0]
We analyze the influence data collection methodology has on two multimodal emotion recognition datasets.
Experiments with the full IEMOCAP dataset indicate that the composition negatively influences generalization performance when compared to the OMG-Emotion Behavior dataset.
arXiv Detail & Related papers (2021-03-05T14:20:59Z) - A Common Operating Picture Framework Leveraging Data Fusion and Deep
Learning [0.7348448478819135]
We present a data fusion framework for accelerating solutions for Processing, Exploitation, and Dissemination.
Our platform is a collection of services that extract information from several data sources by leveraging deep learning and other means of processing.
In our first iteration we have focused on visual data (FMV, WAMI, CCTV/PTZ-Cameras, open source video, etc.) and AIS data streams (satellite and terrestrial sources)
arXiv Detail & Related papers (2020-01-16T18:32:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.