Self-supervised ECG Representation Learning for Emotion Recognition
- URL: http://arxiv.org/abs/2002.03898v2
- Date: Mon, 10 Aug 2020 07:09:41 GMT
- Title: Self-supervised ECG Representation Learning for Emotion Recognition
- Authors: Pritam Sarkar and Ali Etemad
- Abstract summary: We exploit a self-supervised deep multi-task learning framework for electrocardiogram (ECG) -based emotion recognition.
We show that the proposed solution considerably improves the performance compared to a network trained using fully-supervised learning.
- Score: 25.305949034527202
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We exploit a self-supervised deep multi-task learning framework for
electrocardiogram (ECG) -based emotion recognition. The proposed solution
consists of two stages of learning a) learning ECG representations and b)
learning to classify emotions. ECG representations are learned by a signal
transformation recognition network. The network learns high-level abstract
representations from unlabeled ECG data. Six different signal transformations
are applied to the ECG signals, and transformation recognition is performed as
pretext tasks. Training the model on pretext tasks helps the network learn
spatiotemporal representations that generalize well across different datasets
and different emotion categories. We transfer the weights of the
self-supervised network to an emotion recognition network, where the
convolutional layers are kept frozen and the dense layers are trained with
labelled ECG data. We show that the proposed solution considerably improves the
performance compared to a network trained using fully-supervised learning. New
state-of-the-art results are set in classification of arousal, valence,
affective states, and stress for the four utilized datasets. Extensive
experiments are performed, providing interesting insights into the impact of
using a multi-task self-supervised structure instead of a single-task model, as
well as the optimum level of difficulty required for the pretext
self-supervised tasks.
Related papers
- NERULA: A Dual-Pathway Self-Supervised Learning Framework for Electrocardiogram Signal Analysis [5.8961928852930034]
We present NERULA, a self-supervised framework designed for single-lead ECG signals.
NERULA's dual-pathway architecture combines ECG reconstruction and non-contrastive learning to extract detailed cardiac features.
We show that combining generative and discriminative paths into the training spectrum leads to better results by outperforming state-of-the-art self-supervised learning benchmarks in various tasks.
arXiv Detail & Related papers (2024-05-21T14:01:57Z) - Joint Contrastive Learning with Feature Alignment for Cross-Corpus EEG-based Emotion Recognition [2.1645626994550664]
We propose a novel Joint Contrastive learning framework with Feature Alignment to address cross-corpus EEG-based emotion recognition.
In the pre-training stage, a joint domain contrastive learning strategy is introduced to characterize generalizable time-frequency representations of EEG signals.
In the fine-tuning stage, JCFA is refined in conjunction with downstream tasks, where the structural connections among brain electrodes are considered.
arXiv Detail & Related papers (2024-04-15T08:21:17Z) - DGSD: Dynamical Graph Self-Distillation for EEG-Based Auditory Spatial
Attention Detection [49.196182908826565]
Auditory Attention Detection (AAD) aims to detect target speaker from brain signals in a multi-speaker environment.
Current approaches primarily rely on traditional convolutional neural network designed for processing Euclidean data like images.
This paper proposes a dynamical graph self-distillation (DGSD) approach for AAD, which does not require speech stimuli as input.
arXiv Detail & Related papers (2023-09-07T13:43:46Z) - EEG-based Cognitive Load Classification using Feature Masked
Autoencoding and Emotion Transfer Learning [13.404503606887715]
We present a new solution for the classification of cognitive load using electroencephalogram (EEG)
We pre-train our model using self-supervised masked autoencoding on emotion-related EEG datasets.
The results of our experiments show that our proposed approach achieves strong results and outperforms conventional single-stage fully supervised learning.
arXiv Detail & Related papers (2023-08-01T02:59:19Z) - Evaluating the structure of cognitive tasks with transfer learning [67.22168759751541]
This study investigates the transferability of deep learning representations between different EEG decoding tasks.
We conduct extensive experiments using state-of-the-art decoding models on two recently released EEG datasets.
arXiv Detail & Related papers (2023-07-28T14:51:09Z) - GMSS: Graph-Based Multi-Task Self-Supervised Learning for EEG Emotion
Recognition [48.02958969607864]
This paper proposes a graph-based multi-task self-supervised learning model (GMSS) for EEG emotion recognition.
By learning from multiple tasks simultaneously, GMSS can find a representation that captures all of the tasks.
Experiments on SEED, SEED-IV, and MPED datasets show that the proposed model has remarkable advantages in learning more discriminative and general features for EEG emotional signals.
arXiv Detail & Related papers (2022-04-12T03:37:21Z) - Transformer-Based Self-Supervised Learning for Emotion Recognition [0.0]
We propose to use a Transformer-based model to process electrocardiograms (ECG) for emotion recognition.
To overcome the relatively small size of datasets with emotional labels, we employ self-supervised learning.
We show that our approach reaches state-of-the-art performances for emotion recognition using ECG signals on AMIGOS.
arXiv Detail & Related papers (2022-04-08T07:14:55Z) - Learning ECG Representations based on Manipulated Temporal-Spatial
Reverse Detection [11.615287369669971]
We propose a straightforward but effective approach to learn ECG representations.
Inspired by the temporal and spatial characteristics of ECG, we flip the original signals horizontally, vertically, and both horizontally and vertically.
Results show that the ECG representations learned with our method lead to remarkable performances on the downstream task.
arXiv Detail & Related papers (2022-02-25T02:01:09Z) - Improved Speech Emotion Recognition using Transfer Learning and
Spectrogram Augmentation [56.264157127549446]
Speech emotion recognition (SER) is a challenging task that plays a crucial role in natural human-computer interaction.
One of the main challenges in SER is data scarcity.
We propose a transfer learning strategy combined with spectrogram augmentation.
arXiv Detail & Related papers (2021-08-05T10:39:39Z) - Dynamic Graph Modeling of Simultaneous EEG and Eye-tracking Data for
Reading Task Identification [79.41619843969347]
We present a new approach, that we call AdaGTCN, for identifying human reader intent from Electroencephalogram(EEG) and Eye movement(EM) data.
Our method, Adaptive Graph Temporal Convolution Network (AdaGTCN), uses an Adaptive Graph Learning Layer and Deep Neighborhood Graph Convolution Layer.
We compare our approach with several baselines to report an improvement of 6.29% on the ZuCo 2.0 dataset, along with extensive ablation experiments.
arXiv Detail & Related papers (2021-02-21T18:19:49Z) - Semantics-aware Adaptive Knowledge Distillation for Sensor-to-Vision
Action Recognition [131.6328804788164]
We propose a framework, named Semantics-aware Adaptive Knowledge Distillation Networks (SAKDN), to enhance action recognition in vision-sensor modality (videos)
The SAKDN uses multiple wearable-sensors as teacher modalities and uses RGB videos as student modality.
arXiv Detail & Related papers (2020-09-01T03:38:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.