Multi-scale Transformer-based Network for Emotion Recognition from Multi
Physiological Signals
- URL: http://arxiv.org/abs/2305.00769v2
- Date: Mon, 8 May 2023 00:57:02 GMT
- Title: Multi-scale Transformer-based Network for Emotion Recognition from Multi
Physiological Signals
- Authors: Tu Vu and Van Thong Huynh and Soo-Hyung Kim
- Abstract summary: This paper presents an efficient Multi-scale Transformer-based approach for the task of Emotion recognition from Physiological data.
Our approach involves applying a Multi-modal technique combined with scaling data to establish the relationship between internal body signals and human emotions.
Our model achieves decent results on the CASE dataset of the EPiC competition, with an RMSE score of 1.45.
- Score: 11.479653866646762
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents an efficient Multi-scale Transformer-based approach for
the task of Emotion recognition from Physiological data, which has gained
widespread attention in the research community due to the vast amount of
information that can be extracted from these signals using modern sensors and
machine learning techniques. Our approach involves applying a Multi-modal
technique combined with scaling data to establish the relationship between
internal body signals and human emotions. Additionally, we utilize Transformer
and Gaussian Transformation techniques to improve signal encoding effectiveness
and overall performance. Our model achieves decent results on the CASE dataset
of the EPiC competition, with an RMSE score of 1.45.
Related papers
- CNN-Transformer Rectified Collaborative Learning for Medical Image Segmentation [60.08541107831459]
This paper proposes a CNN-Transformer rectified collaborative learning framework to learn stronger CNN-based and Transformer-based models for medical image segmentation.
Specifically, we propose a rectified logit-wise collaborative learning (RLCL) strategy which introduces the ground truth to adaptively select and rectify the wrong regions in student soft labels.
We also propose a class-aware feature-wise collaborative learning (CFCL) strategy to achieve effective knowledge transfer between CNN-based and Transformer-based models in the feature space.
arXiv Detail & Related papers (2024-08-25T01:27:35Z) - BRIEDGE: EEG-Adaptive Edge AI for Multi-Brain to Multi-Robot Interaction [4.815698105652729]
BRIEDGE is an end-to-end system for multi-brain to multi-robot interaction through an EEG-adaptive neural network and an encoding-decoding communication framework.
The encoding-decoding communication framework then encodes the EEG-based semantic information and decodes it into commands in the process of data transmission.
Our experiments show that BRIEDGE achieves the best classification accuracy of heterogeneous EEG data, and more stable performance under noisy environments.
arXiv Detail & Related papers (2024-03-14T15:43:48Z) - Physics-informed and Unsupervised Riemannian Domain Adaptation for Machine Learning on Heterogeneous EEG Datasets [53.367212596352324]
We propose an unsupervised approach leveraging EEG signal physics.
We map EEG channels to fixed positions using field, source-free domain adaptation.
Our method demonstrates robust performance in brain-computer interface (BCI) tasks and potential biomarker applications.
arXiv Detail & Related papers (2024-03-07T16:17:33Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - Convolutional Monge Mapping Normalization for learning on sleep data [63.22081662149488]
We propose a new method called Convolutional Monge Mapping Normalization (CMMN)
CMMN consists in filtering the signals in order to adapt their power spectrum density (PSD) to a Wasserstein barycenter estimated on training data.
Numerical experiments on sleep EEG data show that CMMN leads to significant and consistent performance gains independent from the neural network architecture.
arXiv Detail & Related papers (2023-05-30T08:24:01Z) - CIT-EmotionNet: CNN Interactive Transformer Network for EEG Emotion
Recognition [6.208851183775046]
We propose a novel CNN Interactive Transformer Network for EEG Emotion Recognition, known as CIT-EmotionNet.
We convert raw EEG signals into spatial-frequency representations, which serve as inputs. Then, we integrate Convolutional Neural Network (CNN) and Transformer within a single framework in a parallel manner.
The proposed CIT-EmotionNet outperforms state-of-the-art methods, achieving an average recognition accuracy of 98.57% and 92.09% on two publicly available datasets.
arXiv Detail & Related papers (2023-05-07T16:27:09Z) - Self-Supervised Multimodal Fusion Transformer for Passive Activity
Recognition [2.35066982314539]
Wi-Fi signals provide significant opportunities for human sensing and activity recognition in fields such as healthcare.
Current systems do not effectively exploit the information acquired through multiple sensors to recognise the different activities.
We propose the Fusion Transformer, an attention-based model for multimodal and multi-sensor fusion.
arXiv Detail & Related papers (2022-08-15T15:38:10Z) - A Hybrid Complex-valued Neural Network Framework with Applications to
Electroencephalogram (EEG) [16.578387039499386]
The proposed neural network architecture consists of one complex-valued convolutional layer, two real-valued convolutional layers, and three fully connected layers.
We validate our approach using two simulated EEG signals and a benchmark data set and compare it with two widely used frameworks.
arXiv Detail & Related papers (2022-07-28T00:51:07Z) - Transformer-Based Self-Supervised Learning for Emotion Recognition [0.0]
We propose to use a Transformer-based model to process electrocardiograms (ECG) for emotion recognition.
To overcome the relatively small size of datasets with emotional labels, we employ self-supervised learning.
We show that our approach reaches state-of-the-art performances for emotion recognition using ECG signals on AMIGOS.
arXiv Detail & Related papers (2022-04-08T07:14:55Z) - Multimodal Emotion Recognition using Transfer Learning from Speaker
Recognition and BERT-based models [53.31917090073727]
We propose a neural network-based emotion recognition framework that uses a late fusion of transfer-learned and fine-tuned models from speech and text modalities.
We evaluate the effectiveness of our proposed multimodal approach on the interactive emotional dyadic motion capture dataset.
arXiv Detail & Related papers (2022-02-16T00:23:42Z) - Improved Speech Emotion Recognition using Transfer Learning and
Spectrogram Augmentation [56.264157127549446]
Speech emotion recognition (SER) is a challenging task that plays a crucial role in natural human-computer interaction.
One of the main challenges in SER is data scarcity.
We propose a transfer learning strategy combined with spectrogram augmentation.
arXiv Detail & Related papers (2021-08-05T10:39:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.