A Composable Channel-Adaptive Architecture for Seizure Classification
- URL: http://arxiv.org/abs/2512.19123v1
- Date: Mon, 22 Dec 2025 07:57:20 GMT
- Title: A Composable Channel-Adaptive Architecture for Seizure Classification
- Authors: Francesco Carzaniga, Michael Hersche, Kaspar Schindler, Abbas Rahimi,
- Abstract summary: We develop a channel-adaptive (CA) architecture that seamlessly processes time-series with an arbitrary number of channels.<n>We evaluate our CA-models on a seizure detection task both on a short-term (20 hours) and a long-term (2500 hours) dataset.
- Score: 7.0214066512269
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Objective: We develop a channel-adaptive (CA) architecture that seamlessly processes multi-variate time-series with an arbitrary number of channels, and in particular intracranial electroencephalography (iEEG) recordings. Methods: Our CA architecture first processes the iEEG signal using state-of-the-art models applied to each single channel independently. The resulting features are then fused using a vector-symbolic algorithm which reconstructs the spatial relationship using a trainable scalar per channel. Finally, the fused features are accumulated in a long-term memory of up to 2 minutes to perform the classification. Each CA-model can then be pre-trained on a large corpus of iEEG recordings from multiple heterogeneous subjects. The pre-trained model is personalized to each subject via a quick fine-tuning routine, which uses equal or lower amounts of data compared to existing state-of-the-art models, but requiring only 1/5 of the time. Results: We evaluate our CA-models on a seizure detection task both on a short-term (~20 hours) and a long-term (~2500 hours) dataset. In particular, our CA-EEGWaveNet is trained on a single seizure of the tested subject, while the baseline EEGWaveNet is trained on all but one. Even in this challenging scenario, our CA-EEGWaveNet surpasses the baseline in median F1-score (0.78 vs 0.76). Similarly, CA-EEGNet based on EEGNet, also surpasses its baseline in median F1-score (0.79 vs 0.74). Conclusion and significance: Our CA-model addresses two issues: first, it is channel-adaptive and can therefore be trained across heterogeneous subjects without loss of performance; second, it increases the effective temporal context size to a clinically-relevant length. Therefore, our model is a drop-in replacement for existing models, bringing better characteristics and performance across the board.
Related papers
- WaveNet's Precision in EEG Classification [1.0885910878567457]
This study introduces a WaveNet-based deep learning model designed to automate the classification of EEG signals into physiological, pathological, artifact, and noise categories.<n>The model was trained, validated, and tested on 209,232 samples with a 70/20/10 percent split.<n>WaveNet's architecture, originally developed for raw audio synthesis, is well suited for EEG data due to its use of dilated causal convolutions and residual connections.
arXiv Detail & Related papers (2025-10-10T09:21:21Z) - ALFEE: Adaptive Large Foundation Model for EEG Representation [17.166788472910806]
We propose the Adaptive Large Foundation model for EEG signal representation(ALFEE) framework.<n>ALFEE is a novel hybrid transformer architecture with two learning stages for robust EEG representation learning.<n>After 25,000 hours of pretraining, extensive experimental results on six downstream EEG tasks demonstrate the superior performance of ALFEE over existing models.
arXiv Detail & Related papers (2025-05-07T13:32:31Z) - WaveGNN: Modeling Irregular Multivariate Time Series for Accurate Predictions [3.489870763747715]
Real-world time series often exhibit irregularities such as misaligned timestamps, missing entries, and variable sampling rates.<n>Existing approaches often rely on imputation, which can introduce biases.<n>We present WaveGNN, a novel framework designed to embed irregularly sampled time series data for accurate predictions.
arXiv Detail & Related papers (2024-12-14T00:03:44Z) - Consistency Trajectory Models: Learning Probability Flow ODE Trajectory of Diffusion [56.38386580040991]
Consistency Trajectory Model (CTM) is a generalization of Consistency Models (CM)
CTM enables the efficient combination of adversarial training and denoising score matching loss to enhance performance.
Unlike CM, CTM's access to the score function can streamline the adoption of established controllable/conditional generation methods.
arXiv Detail & Related papers (2023-10-01T05:07:17Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - Your representations are in the network: composable and parallel
adaptation for large scale models [90.26965623489157]
InCA is a lightweight method for transfer learning that cross-attends to any activation layer of a pre-trained model.
We show that, even when selecting a single top-scoring adapter, InCA achieves performance comparable to full fine-tuning.
arXiv Detail & Related papers (2023-03-07T18:12:24Z) - HARDC : A novel ECG-based heartbeat classification method to detect
arrhythmia using hierarchical attention based dual structured RNN with
dilated CNN [3.8791511769387625]
We have developed a novel hybrid hierarchical attention-based bidirectional recurrent neural network with dilated CNN (HARDC) method for arrhythmia classification.
The proposed HARDC fully exploits the dilated CNN and bidirectional recurrent neural network unit (BiGRU-BiLSTM) architecture to generate fusion features.
Our results indicate that an automated and highly computed method to classify multiple types of arrhythmia signals holds considerable promise.
arXiv Detail & Related papers (2023-03-06T13:26:29Z) - Prompt Tuning for Parameter-efficient Medical Image Segmentation [79.09285179181225]
We propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets.
We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes.
We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models.
arXiv Detail & Related papers (2022-11-16T21:55:05Z) - Layer Pruning on Demand with Intermediate CTC [50.509073206630994]
We present a training and pruning method for ASR based on the connectionist temporal classification (CTC)
We show that a Transformer-CTC model can be pruned in various depth on demand, improving real-time factor from 0.005 to 0.002 on GPU.
arXiv Detail & Related papers (2021-06-17T02:40:18Z) - Deep transfer learning for improving single-EEG arousal detection [63.52264764099532]
Two datasets do not contain exactly the same setup leading to degraded performance in single-EEG models.
We train a baseline model and replace the first two layers to prepare the architecture for single-channel electroencephalography data.
Using a fine-tuning strategy, our model yields similar performance to the baseline model and was significantly better than a comparable single-channel model.
arXiv Detail & Related papers (2020-04-10T16:51:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.