Deep learning for ECoG brain-computer interface: end-to-end vs.
hand-crafted features
- URL: http://arxiv.org/abs/2210.02544v1
- Date: Wed, 5 Oct 2022 20:18:30 GMT
- Title: Deep learning for ECoG brain-computer interface: end-to-end vs.
hand-crafted features
- Authors: Maciej \'Sliwowski, Matthieu Martin, Antoine Souloumiac, Pierre
Blanchart, Tetiana Aksenova
- Abstract summary: Brain signals are temporal data with a low signal-to-noise ratio, uncertain labels, and nonstationary data in time.
These factors may influence the training process and slow down the models' performance improvement.
This paper compares models that use raw ECoG signal and time-frequency features for BCI motor imagery decoding.
- Score: 4.7773230870500605
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In brain signal processing, deep learning (DL) models have become commonly
used. However, the performance gain from using end-to-end DL models compared to
conventional ML approaches is usually significant but moderate, typically at
the cost of increased computational load and deteriorated explainability. The
core idea behind deep learning approaches is scaling the performance with
bigger datasets. However, brain signals are temporal data with a low
signal-to-noise ratio, uncertain labels, and nonstationary data in time. Those
factors may influence the training process and slow down the models'
performance improvement. These factors' influence may differ for end-to-end DL
model and one using hand-crafted features. As not studied before, this paper
compares models that use raw ECoG signal and time-frequency features for BCI
motor imagery decoding. We investigate whether the current dataset size is a
stronger limitation for any models. Finally, obtained filters were compared to
identify differences between hand-crafted features and optimized with
backpropagation. To compare the effectiveness of both strategies, we used a
multilayer perceptron and a mix of convolutional and LSTM layers that were
already proved effective in this task. The analysis was performed on the
long-term clinical trial database (almost 600 minutes of recordings) of a
tetraplegic patient executing motor imagery tasks for 3D hand translation. For
a given dataset, the results showed that end-to-end training might not be
significantly better than the hand-crafted features-based model. The
performance gap is reduced with bigger datasets, but considering the increased
computational load, end-to-end training may not be profitable for this
application.
Related papers
- Evolutionary Retrofitting [42.21143557577615]
AfterLearnER consists in applying non-differentiable optimization, including evolutionary methods, to fully-trained machine learning models.
The efficiency of AfterLearnER is demonstrated by tackling non-differentiable signals such as threshold-based criteria in depth sensing, the word error rate in speech re-synthesis, image quality in 3D generative adversarial networks (GANs)
The advantages of AfterLearnER are its versatility (no gradient is needed), the possibility to use non-differentiable feedback including human evaluations, the limited overfitting, supported by a theoretical study and its anytime behavior.
arXiv Detail & Related papers (2024-10-15T06:59:32Z) - Reducing Data Bottlenecks in Distributed, Heterogeneous Neural Networks [5.32129361961937]
This paper investigates the impact of bottleneck size on the performance of deep learning models in embedded multicore and many-core systems.
We apply a hardware-software co-design methodology where data bottlenecks are replaced with extremely narrow layers to reduce the amount of data traffic.
Hardware-side evaluation reveals that higher bottleneck ratios lead to substantial reductions in data transfer volume across the layers of the neural network.
arXiv Detail & Related papers (2024-10-12T21:07:55Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Impact of dataset size and long-term ECoG-based BCI usage on deep
learning decoders performance [4.7773230870500605]
In brain-computer interfaces (BCI) research, recording data is time-consuming and expensive.
Can we achieve higher decoding performance with more data to train decoders?
High decoding performance was obtained with relatively small datasets recorded later in the experiment.
arXiv Detail & Related papers (2022-09-08T13:01:05Z) - Decoding ECoG signal into 3D hand translation using deep learning [3.20238141000059]
Motor brain-computer interfaces (BCIs) are promising technology that may enable motor-impaired people to interact with their environment.
Most ECoG signal decoders used to predict continuous hand movements are linear models.
Deep learning models, which are state-of-the-art in many problems, could be a solution to better capture this relationship.
arXiv Detail & Related papers (2021-10-05T15:41:04Z) - The Devil Is in the Details: An Efficient Convolutional Neural Network
for Transport Mode Detection [3.008051369744002]
Transport mode detection is a classification problem aiming to design an algorithm that can infer the transport mode of a user given multimodal signals.
We show that a small, optimized model can perform as well as a current deep model.
arXiv Detail & Related papers (2021-09-16T08:05:47Z) - STAR: Sparse Transformer-based Action Recognition [61.490243467748314]
This work proposes a novel skeleton-based human action recognition model with sparse attention on the spatial dimension and segmented linear attention on the temporal dimension of data.
Experiments show that our model can achieve comparable performance while utilizing much less trainable parameters and achieve high speed in training and inference.
arXiv Detail & Related papers (2021-07-15T02:53:11Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Convolutional Tensor-Train LSTM for Spatio-temporal Learning [116.24172387469994]
We propose a higher-order LSTM model that can efficiently learn long-term correlations in the video sequence.
This is accomplished through a novel tensor train module that performs prediction by combining convolutional features across time.
Our results achieve state-of-the-art performance-art in a wide range of applications and datasets.
arXiv Detail & Related papers (2020-02-21T05:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.