Neuro-GPT: Towards A Foundation Model for EEG
- URL: http://arxiv.org/abs/2311.03764v4
- Date: Sat, 2 Mar 2024 07:06:39 GMT
- Title: Neuro-GPT: Towards A Foundation Model for EEG
- Authors: Wenhui Cui, Woojae Jeong, Philipp Th\"olke, Takfarinas Medani, Karim
Jerbi, Anand A. Joshi, Richard M. Leahy
- Abstract summary: We propose Neuro-GPT, a foundation model consisting of an EEG encoder and a GPT model.
Foundation model is pre-trained on a large-scale data set using a self-supervised task that learns how to reconstruct masked EEG segments.
Experiments demonstrate that applying a foundation model can significantly improve classification performance compared to a model trained from scratch.
- Score: 0.04188114563181615
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: To handle the scarcity and heterogeneity of electroencephalography (EEG) data
for Brain-Computer Interface (BCI) tasks, and to harness the power of large
publicly available data sets, we propose Neuro-GPT, a foundation model
consisting of an EEG encoder and a GPT model. The foundation model is
pre-trained on a large-scale data set using a self-supervised task that learns
how to reconstruct masked EEG segments. We then fine-tune the model on a Motor
Imagery Classification task to validate its performance in a low-data regime (9
subjects). Our experiments demonstrate that applying a foundation model can
significantly improve classification performance compared to a model trained
from scratch, which provides evidence for the generalizability of the
foundation model and its ability to address challenges of data scarcity and
heterogeneity in EEG. The code is publicly available at
github.com/wenhui0206/NeuroGPT.
Related papers
- Graph Adapter of EEG Foundation Models for Parameter Efficient Fine Tuning [1.8946099300030472]
EEG-GraphAdapter (EGA) is a parameter-efficient fine-tuning (PEFT) approach to address these challenges.
EGA is integrated into pre-trained temporal backbone models as a GNN-based module.
It improves performance by up to 16.1% in the F1-score compared with the backbone BENDR model.
arXiv Detail & Related papers (2024-11-25T07:30:52Z) - EEGPT: Unleashing the Potential of EEG Generalist Foundation Model by Autoregressive Pre-training [9.57946371147345]
EEGPT is the first generalist EEG foundation model designed to address these challenges.
First, we propose an electrode-wise modeling strategy that treats each electrode as a fundamental unit.
Second, we develop the first autoregressive EEG pre-trained model.
Third, we introduce a multi-task transfer learning paradigm using a learnable electrode graph network.
arXiv Detail & Related papers (2024-10-14T12:17:54Z) - EEGFormer: Towards Transferable and Interpretable Large-Scale EEG
Foundation Model [39.363511340878624]
We present a novel EEG foundation model, namely EEGFormer, pretrained on large-scale compound EEG data.
To validate the effectiveness of our model, we extensively evaluate it on various downstream tasks and assess the performance under different transfer settings.
arXiv Detail & Related papers (2024-01-11T17:36:24Z) - hvEEGNet: exploiting hierarchical VAEs on EEG data for neuroscience
applications [3.031375888004876]
Two main issues challenge the existing DL-based modeling methods for EEG.
High variability between subjects and low signal-to-noise ratio make it difficult to ensure a good quality in the EEG data.
We propose two variational autoencoder models, namely vEEGNet-ver3 and hvEEGNet, to target the problem of high-fidelity EEG reconstruction.
arXiv Detail & Related papers (2023-11-20T15:36:31Z) - Universal Domain Adaptation from Foundation Models: A Baseline Study [58.51162198585434]
We make empirical studies of state-of-the-art UniDA methods using foundation models.
We introduce textitCLIP distillation, a parameter-free method specifically designed to distill target knowledge from CLIP models.
Although simple, our method outperforms previous approaches in most benchmark tasks.
arXiv Detail & Related papers (2023-05-18T16:28:29Z) - Towards Efficient Task-Driven Model Reprogramming with Foundation Models [52.411508216448716]
Vision foundation models exhibit impressive power, benefiting from the extremely large model capacity and broad training data.
However, in practice, downstream scenarios may only support a small model due to the limited computational resources or efficiency considerations.
This brings a critical challenge for the real-world application of foundation models: one has to transfer the knowledge of a foundation model to the downstream task.
arXiv Detail & Related papers (2023-04-05T07:28:33Z) - Towards physiology-informed data augmentation for EEG-based BCIs [24.15108821320151]
We suggest a novel technique for augmenting the training data by generating new data from the data set at hand.
In this manuscript, we explain the method and show first preliminary results for participant-independent motor-imagery classification.
arXiv Detail & Related papers (2022-03-27T20:59:40Z) - A multi-stage machine learning model on diagnosis of esophageal
manometry [50.591267188664666]
The framework includes deep-learning models at the swallow-level stage and feature-based machine learning models at the study-level stage.
This is the first artificial-intelligence-style model to automatically predict CC diagnosis of HRM study from raw multi-swallow data.
arXiv Detail & Related papers (2021-06-25T20:09:23Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z) - CHEER: Rich Model Helps Poor Model via Knowledge Infusion [69.23072792708263]
We develop a knowledge infusion framework named CHEER that can succinctly summarize such rich model into transferable representations.
Our empirical results showed that CHEER outperformed baselines by 5.60% to 46.80% in terms of the macro-F1 score on multiple physiological datasets.
arXiv Detail & Related papers (2020-05-21T21:44:21Z) - Deep transfer learning for improving single-EEG arousal detection [63.52264764099532]
Two datasets do not contain exactly the same setup leading to degraded performance in single-EEG models.
We train a baseline model and replace the first two layers to prepare the architecture for single-channel electroencephalography data.
Using a fine-tuning strategy, our model yields similar performance to the baseline model and was significantly better than a comparable single-channel model.
arXiv Detail & Related papers (2020-04-10T16:51:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.