Transformer-based Spatial-Temporal Feature Learning for EEG Decoding
- URL: http://arxiv.org/abs/2106.11170v1
- Date: Fri, 11 Jun 2021 00:48:18 GMT
- Title: Transformer-based Spatial-Temporal Feature Learning for EEG Decoding
- Authors: Yonghao Song, Xueyu Jia, Lie Yang, Longhan Xie
- Abstract summary: We propose a novel EEG decoding method that mainly relies on the attention mechanism.
We have reached the level of the state-of-the-art in multi-classification of EEG, with fewer parameters.
It has good potential to promote the practicality of brain-computer interface (BCI)
- Score: 4.8276709243429
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: At present, people usually use some methods based on convolutional neural
networks (CNNs) for Electroencephalograph (EEG) decoding. However, CNNs have
limitations in perceiving global dependencies, which is not adequate for common
EEG paradigms with a strong overall relationship. Regarding this issue, we
propose a novel EEG decoding method that mainly relies on the attention
mechanism. The EEG data is firstly preprocessed and spatially filtered. And
then, we apply attention transforming on the feature-channel dimension so that
the model can enhance more relevant spatial features. The most crucial step is
to slice the data in the time dimension for attention transforming, and finally
obtain a highly distinguishable representation. At this time, global averaging
pooling and a simple fully-connected layer are used to classify different
categories of EEG data. Experiments on two public datasets indicate that the
strategy of attention transforming effectively utilizes spatial and temporal
features. And we have reached the level of the state-of-the-art in
multi-classification of EEG, with fewer parameters. As far as we know, it is
the first time that a detailed and complete method based on the transformer
idea has been proposed in this field. It has good potential to promote the
practicality of brain-computer interface (BCI). The source code can be found
at: \textit{https://github.com/anranknight/EEG-Transformer}.
Related papers
- How Homogenizing the Channel-wise Magnitude Can Enhance EEG Classification Model? [4.0871083166108395]
We propose a simple yet effective approach for EEG data pre-processing.
Our method first transforms the EEG data into an encoded image by an Inverted Channel-wise Magnitude Homogenization.
By doing so, we can improve the EEG learning process efficiently without using a huge Deep Learning network.
arXiv Detail & Related papers (2024-07-19T09:11:56Z) - Large Brain Model for Learning Generic Representations with Tremendous EEG Data in BCI [6.926908480247951]
We propose a unified foundation model for EEG called Large Brain Model (LaBraM)
LaBraM enables cross-dataset learning by segmenting the EEG signals into EEG channel patches.
We then pre-train neural Transformers by predicting the original neural codes for the masked EEG channel patches.
arXiv Detail & Related papers (2024-05-29T05:08:16Z) - EEG-Deformer: A Dense Convolutional Transformer for Brain-computer Interfaces [17.524441950422627]
We introduce EEG-Deformer, which incorporates two main novel components into a CNN-Transformer.
EEG-Deformer learns from neurophysiologically meaningful brain regions for the corresponding cognitive tasks.
arXiv Detail & Related papers (2024-04-25T18:00:46Z) - EEG-ITNet: An Explainable Inception Temporal Convolutional Network for
Motor Imagery Classification [0.5616884466478884]
We propose an end-to-end deep learning architecture called EEG-ITNet.
Our model can extract rich spectral, spatial, and temporal information from multi-channel EEG signals.
EEG-ITNet shows up to 5.9% improvement in the classification accuracy in different scenarios.
arXiv Detail & Related papers (2022-04-14T13:18:43Z) - Learning A 3D-CNN and Transformer Prior for Hyperspectral Image
Super-Resolution [80.93870349019332]
We propose a novel HSISR method that uses Transformer instead of CNN to learn the prior of HSIs.
Specifically, we first use the gradient algorithm to solve the HSISR model, and then use an unfolding network to simulate the iterative solution processes.
arXiv Detail & Related papers (2021-11-27T15:38:57Z) - HAT: Hierarchical Aggregation Transformers for Person Re-identification [87.02828084991062]
We take advantages of both CNNs and Transformers for image-based person Re-ID with high performance.
Work is the first to take advantages of both CNNs and Transformers for image-based person Re-ID.
arXiv Detail & Related papers (2021-07-13T09:34:54Z) - CADDA: Class-wise Automatic Differentiable Data Augmentation for EEG
Signals [92.60744099084157]
We propose differentiable data augmentation amenable to gradient-based learning.
We demonstrate the relevance of our approach on the clinically relevant sleep staging classification task.
arXiv Detail & Related papers (2021-06-25T15:28:48Z) - TELESTO: A Graph Neural Network Model for Anomaly Classification in
Cloud Services [77.454688257702]
Machine learning (ML) and artificial intelligence (AI) are applied on IT system operation and maintenance.
One direction aims at the recognition of re-occurring anomaly types to enable remediation automation.
We propose a method that is invariant to dimensionality changes of given data.
arXiv Detail & Related papers (2021-02-25T14:24:49Z) - EEG-Inception: An Accurate and Robust End-to-End Neural Network for
EEG-based Motor Imagery Classification [123.93460670568554]
This paper proposes a novel convolutional neural network (CNN) architecture for accurate and robust EEG-based motor imagery (MI) classification.
The proposed CNN model, namely EEG-Inception, is built on the backbone of the Inception-Time network.
The proposed network is an end-to-end classification, as it takes the raw EEG signals as the input and does not require complex EEG signal-preprocessing.
arXiv Detail & Related papers (2021-01-24T19:03:10Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - EEG Channel Interpolation Using Deep Encoder-decoder Netwoks [6.420508750179696]
"Pop" artifacts originate from the spontaneous loss of connectivity between a surface and an electrode.
EEG uses a dense array of electrodes, hence "popped" segments are among the most pervasive type of artifact seen during the collection of EEG data.
In this paper we frame the brain problem as a self-learning task using a deep encoder-decoder network.
arXiv Detail & Related papers (2020-09-21T21:57:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.