Relate auditory speech to EEG by shallow-deep attention-based network
- URL: http://arxiv.org/abs/2303.10897v1
- Date: Mon, 20 Mar 2023 06:34:22 GMT
- Title: Relate auditory speech to EEG by shallow-deep attention-based network
- Authors: Fan Cui, Liyong Guo, Lang He, Jiyao Liu, ErCheng Pei, Yujun Wang,
Dongmei Jiang
- Abstract summary: We propose a novel Shallow-Deep Attention-based Network (SDANet) to classify the correct auditory stimulus evoking the EEG signal.
It adopts the Attention-based Correlation Module (ACM) to discover the connection between auditory speech and EEG from global aspect.
Various training strategies and data augmentation are used to boost the model robustness.
- Score: 10.002888298492831
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Electroencephalography (EEG) plays a vital role in detecting how brain
responses to different stimulus. In this paper, we propose a novel Shallow-Deep
Attention-based Network (SDANet) to classify the correct auditory stimulus
evoking the EEG signal. It adopts the Attention-based Correlation Module (ACM)
to discover the connection between auditory speech and EEG from global aspect,
and the Shallow-Deep Similarity Classification Module (SDSCM) to decide the
classification result via the embeddings learned from the shallow and deep
layers. Moreover, various training strategies and data augmentation are used to
boost the model robustness. Experiments are conducted on the dataset provided
by Auditory EEG challenge (ICASSP Signal Processing Grand Challenge 2023).
Results show that the proposed model has a significant gain over the baseline
on the match-mismatch track.
Related papers
- Single-word Auditory Attention Decoding Using Deep Learning Model [9.698931956476692]
Identifying auditory attention by comparing auditory stimuli and corresponding brain responses, is known as auditory attention decoding (AAD)
This paper presents a deep learning approach, based on EEGNet, to address this challenge.
arXiv Detail & Related papers (2024-10-15T21:57:19Z) - NeuroSpex: Neuro-Guided Speaker Extraction with Cross-Modal Attention [47.8479647938849]
We present a neuro-guided speaker extraction model, i.e. NeuroSpex, using the EEG response of the listener as the sole auxiliary reference cue.
We propose a novel EEG signal encoder that captures the attention information. Additionally, we propose a cross-attention (CA) mechanism to enhance the speech feature representations.
arXiv Detail & Related papers (2024-09-04T07:33:01Z) - DGSD: Dynamical Graph Self-Distillation for EEG-Based Auditory Spatial
Attention Detection [49.196182908826565]
Auditory Attention Detection (AAD) aims to detect target speaker from brain signals in a multi-speaker environment.
Current approaches primarily rely on traditional convolutional neural network designed for processing Euclidean data like images.
This paper proposes a dynamical graph self-distillation (DGSD) approach for AAD, which does not require speech stimuli as input.
arXiv Detail & Related papers (2023-09-07T13:43:46Z) - EEGminer: Discovering Interpretable Features of Brain Activity with
Learnable Filters [72.19032452642728]
We propose a novel differentiable EEG decoding pipeline consisting of learnable filters and a pre-determined feature extraction module.
We demonstrate the utility of our model towards emotion recognition from EEG signals on the SEED dataset and on a new EEG dataset of unprecedented size.
The discovered features align with previous neuroscience studies and offer new insights, such as marked differences in the functional connectivity profile between left and right temporal areas during music listening.
arXiv Detail & Related papers (2021-10-19T14:22:04Z) - Improved Speech Emotion Recognition using Transfer Learning and
Spectrogram Augmentation [56.264157127549446]
Speech emotion recognition (SER) is a challenging task that plays a crucial role in natural human-computer interaction.
One of the main challenges in SER is data scarcity.
We propose a transfer learning strategy combined with spectrogram augmentation.
arXiv Detail & Related papers (2021-08-05T10:39:39Z) - EEG-based Cross-Subject Driver Drowsiness Recognition with an
Interpretable Convolutional Neural Network [0.0]
We develop a novel convolutional neural network combined with an interpretation technique that allows sample-wise analysis of important features for classification.
Results show that the model achieves an average accuracy of 78.35% on 11 subjects for leave-one-out cross-subject recognition.
arXiv Detail & Related papers (2021-05-30T14:47:20Z) - Discriminative Singular Spectrum Classifier with Applications on
Bioacoustic Signal Recognition [67.4171845020675]
We present a bioacoustic signal classifier equipped with a discriminative mechanism to extract useful features for analysis and classification efficiently.
Unlike current bioacoustic recognition methods, which are task-oriented, the proposed model relies on transforming the input signals into vector subspaces.
The validity of the proposed method is verified using three challenging bioacoustic datasets containing anuran, bee, and mosquito species.
arXiv Detail & Related papers (2021-03-18T11:01:21Z) - Emotional EEG Classification using Connectivity Features and
Convolutional Neural Networks [81.74442855155843]
We introduce a new classification system that utilizes brain connectivity with a CNN and validate its effectiveness via the emotional video classification.
The level of concentration of the brain connectivity related to the emotional property of the target video is correlated with classification performance.
arXiv Detail & Related papers (2021-01-18T13:28:08Z) - Multi-Lead ECG Classification via an Information-Based Attention
Convolutional Neural Network [1.1720399305661802]
One-dimensional convolutional neural networks (CNN) have proven to be effective in pervasive classification tasks.
We implement the Residual connection and design a structure which can learn the weights from the information contained in different channels in the input feature map.
An indicator named mean square deviation is introduced to monitor the performance of a particular model segment in the classification task.
arXiv Detail & Related papers (2020-03-25T02:28:04Z) - Multi-Scale Neural network for EEG Representation Learning in BCI [2.105172041656126]
We propose a novel deep multi-scale neural network that discovers feature representations in multiple frequency/time ranges.
By representing EEG signals withspectral-temporal information, the proposed method can be utilized for diverse paradigms.
arXiv Detail & Related papers (2020-03-02T04:06:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.