TSception: A Deep Learning Framework for Emotion Detection Using EEG
- URL: http://arxiv.org/abs/2004.02965v2
- Date: Wed, 8 Apr 2020 01:39:59 GMT
- Title: TSception: A Deep Learning Framework for Emotion Detection Using EEG
- Authors: Yi Ding, Neethu Robinson, Qiuhao Zeng, Duo Chen, Aung Aung Phyo Wai,
Tih-Shih Lee, Cuntai Guan
- Abstract summary: We propose a deep learning framework, TSception, for emotion detection from electroencephalogram (EEG)
TSception consists of temporal and spatial convolutional layers, which learn discriminative representations in the time and channel domains simultaneously.
TSception achieves a high classification accuracy of 86.03%, which outperforms the prior methods significantly.
- Score: 11.444502210936776
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a deep learning framework, TSception, for emotion
detection from electroencephalogram (EEG). TSception consists of temporal and
spatial convolutional layers, which learn discriminative representations in the
time and channel domains simultaneously. The temporal learner consists of
multi-scale 1D convolutional kernels whose lengths are related to the sampling
rate of the EEG signal, which learns multiple temporal and frequency
representations. The spatial learner takes advantage of the asymmetry property
of emotion responses at the frontal brain area to learn the discriminative
representations from the left and right hemispheres of the brain. In our study,
a system is designed to study the emotional arousal in an immersive virtual
reality (VR) environment. EEG data were collected from 18 healthy subjects
using this system to evaluate the performance of the proposed deep learning
network for the classification of low and high emotional arousal states. The
proposed method is compared with SVM, EEGNet, and LSTM. TSception achieves a
high classification accuracy of 86.03%, which outperforms the prior methods
significantly (p<0.05). The code is available at
https://github.com/deepBrains/TSception
Related papers
- EGNN-C+: Interpretable Evolving Granular Neural Network and Application
in Classification of Weakly-Supervised EEG Data Streams [0.0]
We introduce a modified incremental learning algorithm for evolving Granular Neural Networks (eGNN-C+)
We focus on the classification of emotion-related patterns within electroencephalogram (EEG) signals.
arXiv Detail & Related papers (2024-02-26T15:11:41Z) - MASA-TCN: Multi-anchor Space-aware Temporal Convolutional Neural
Networks for Continuous and Discrete EEG Emotion Recognition [11.882642356358883]
We propose a novel model, named MASA-TCN, for EEG emotion regression and classification tasks.
The space-aware temporal layer enables TCN to additionally learn from spatial relations among EEG electrodes.
Experiments show MASA-TCN achieves higher results than the state-of-the-art methods for both EEG emotion regression and classification tasks.
arXiv Detail & Related papers (2023-08-30T04:49:24Z) - A Hierarchical Regression Chain Framework for Affective Vocal Burst
Recognition [72.36055502078193]
We propose a hierarchical framework, based on chain regression models, for affective recognition from vocal bursts.
To address the challenge of data sparsity, we also use self-supervised learning (SSL) representations with layer-wise and temporal aggregation modules.
The proposed systems participated in the ACII Affective Vocal Burst (A-VB) Challenge 2022 and ranked first in the "TWO'' and "CULTURE" tasks.
arXiv Detail & Related papers (2023-03-14T16:08:45Z) - fMRI from EEG is only Deep Learning away: the use of interpretable DL to
unravel EEG-fMRI relationships [68.8204255655161]
We present an interpretable domain grounded solution to recover the activity of several subcortical regions from multichannel EEG data.
We recover individual spatial and time-frequency patterns of scalp EEG predictive of the hemodynamic signal in the subcortical nuclei.
arXiv Detail & Related papers (2022-10-23T15:11:37Z) - EEGminer: Discovering Interpretable Features of Brain Activity with
Learnable Filters [72.19032452642728]
We propose a novel differentiable EEG decoding pipeline consisting of learnable filters and a pre-determined feature extraction module.
We demonstrate the utility of our model towards emotion recognition from EEG signals on the SEED dataset and on a new EEG dataset of unprecedented size.
The discovered features align with previous neuroscience studies and offer new insights, such as marked differences in the functional connectivity profile between left and right temporal areas during music listening.
arXiv Detail & Related papers (2021-10-19T14:22:04Z) - Contrastive Learning of Subject-Invariant EEG Representations for
Cross-Subject Emotion Recognition [9.07006689672858]
We propose Contrast Learning method for Inter-Subject Alignment (ISA) for reliable cross-subject emotion recognition.
ISA involves maximizing the similarity in EEG signals across subjects when they received the same stimuli in contrast to different ones.
A convolutional neural network with depthwise spatial convolution and temporal convolution layers was applied to learn inter-subject representations from raw EEG signals.
arXiv Detail & Related papers (2021-09-20T14:13:45Z) - SFE-Net: EEG-based Emotion Recognition with Symmetrical Spatial Feature
Extraction [1.8047694351309205]
We present a spatial folding ensemble network (SFENet) for EEG feature extraction and emotion recognition.
Motivated by the spatial symmetry mechanism of human brain, we fold the input EEG channel data with five different symmetrical strategies.
With this network, the spatial features of different symmetric folding signlas can be extracted simultaneously, which greatly improves the robustness and accuracy of feature recognition.
arXiv Detail & Related papers (2021-04-09T12:59:38Z) - TSception: Capturing Temporal Dynamics and Spatial Asymmetry from EEG
for Emotion Recognition [9.825158483198113]
TSception is a multi-scale convolutional neural network to learn temporal dynamics from affective electroencephalogram (EEG)
The proposed method can be utilized in emotion regulation therapy for emotion recognition in the future.
arXiv Detail & Related papers (2021-04-07T06:10:01Z) - Emotional EEG Classification using Connectivity Features and
Convolutional Neural Networks [81.74442855155843]
We introduce a new classification system that utilizes brain connectivity with a CNN and validate its effectiveness via the emotional video classification.
The level of concentration of the brain connectivity related to the emotional property of the target video is correlated with classification performance.
arXiv Detail & Related papers (2021-01-18T13:28:08Z) - An End-to-End Visual-Audio Attention Network for Emotion Recognition in
User-Generated Videos [64.91614454412257]
We propose to recognize video emotions in an end-to-end manner based on convolutional neural networks (CNNs)
Specifically, we develop a deep Visual-Audio Attention Network (VAANet), a novel architecture that integrates spatial, channel-wise, and temporal attentions into a visual 3D CNN and temporal attentions into an audio 2D CNN.
arXiv Detail & Related papers (2020-02-12T15:33:59Z) - Continuous Emotion Recognition via Deep Convolutional Autoencoder and
Support Vector Regressor [70.2226417364135]
It is crucial that the machine should be able to recognize the emotional state of the user with high accuracy.
Deep neural networks have been used with great success in recognizing emotions.
We present a new model for continuous emotion recognition based on facial expression recognition.
arXiv Detail & Related papers (2020-01-31T17:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.