Do Not Sleep on Linear Models: Simple and Interpretable Techniques
Outperform Deep Learning for Sleep Scoring
- URL: http://arxiv.org/abs/2207.07753v2
- Date: Tue, 19 Jul 2022 12:56:54 GMT
- Title: Do Not Sleep on Linear Models: Simple and Interpretable Techniques
Outperform Deep Learning for Sleep Scoring
- Authors: Jeroen Van Der Donckt, Jonas Van Der Donckt, Emiel Deprost, Nicolas
Vandenbussche, Michael Rademaker, Gilles Vandewiele, Sofie Van Hoecke
- Abstract summary: We argue that most deep learning solutions for sleep scoring are limited in their real-world applicability as they are hard to train, deploy, and reproduce.
In this work, we revisit the problem of sleep stage classification using classical machine learning.
Results show that state-of-the-art performance can be achieved with a conventional machine learning pipeline.
- Score: 1.6339105551302067
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over the last few years, research in automatic sleep scoring has mainly
focused on developing increasingly complex deep learning architectures.
However, recently these approaches achieved only marginal improvements, often
at the expense of requiring more data and more expensive training procedures.
Despite all these efforts and their satisfactory performance, automatic sleep
staging solutions are not widely adopted in a clinical context yet. We argue
that most deep learning solutions for sleep scoring are limited in their
real-world applicability as they are hard to train, deploy, and reproduce.
Moreover, these solutions lack interpretability and transparency, which are
often key to increase adoption rates. In this work, we revisit the problem of
sleep stage classification using classical machine learning. Results show that
state-of-the-art performance can be achieved with a conventional machine
learning pipeline consisting of preprocessing, feature extraction, and a simple
machine learning model. In particular, we analyze the performance of a linear
model and a non-linear (gradient boosting) model. Our approach surpasses
state-of-the-art (that uses the same data) on two public datasets: Sleep-EDF
SC-20 (MF1 0.810) and Sleep-EDF ST (MF1 0.795), while achieving competitive
results on Sleep-EDF SC-78 (MF1 0.775) and MASS SS3 (MF1 0.817). We show that,
for the sleep stage scoring task, the expressiveness of an engineered feature
vector is on par with the internally learned representations of deep learning
models. This observation opens the door to clinical adoption, as a
representative feature vector allows to leverage both the interpretability and
successful track record of traditional machine learning models.
Related papers
- PILOT: A Pre-Trained Model-Based Continual Learning Toolbox [71.63186089279218]
This paper introduces a pre-trained model-based continual learning toolbox known as PILOT.
On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt.
On the other hand, PILOT fits typical class-incremental learning algorithms within the context of pre-trained models to evaluate their effectiveness.
arXiv Detail & Related papers (2023-09-13T17:55:11Z) - Transparency in Sleep Staging: Deep Learning Method for EEG Sleep Stage
Classification with Model Interpretability [5.747465732334616]
This study presents an end-to-end deep learning (DL) model which integrates squeeze and excitation blocks within the residual network to extract features and stacked Bi-LSTM to understand complex temporal dependencies.
A distinctive aspect of this study is the adaptation of GradCam for sleep staging, marking the first instance of an explainable DL model in this domain with alignment of its decision-making with sleep expert's insights.
arXiv Detail & Related papers (2023-09-10T17:56:03Z) - RanPAC: Random Projections and Pre-trained Models for Continual Learning [59.07316955610658]
Continual learning (CL) aims to learn different tasks (such as classification) in a non-stationary data stream without forgetting old ones.
We propose a concise and effective approach for CL with pre-trained models.
arXiv Detail & Related papers (2023-07-05T12:49:02Z) - Phantom Embeddings: Using Embedding Space for Model Regularization in
Deep Neural Networks [12.293294756969477]
The strength of machine learning models stems from their ability to learn complex function approximations from data.
The complex models tend to memorize the training data, which results in poor regularization performance on test data.
We present a novel approach to regularize the models by leveraging the information-rich latent embeddings and their high intra-class correlation.
arXiv Detail & Related papers (2023-04-14T17:15:54Z) - A CNN-Transformer Deep Learning Model for Real-time Sleep Stage
Classification in an Energy-Constrained Wireless Device [2.5672176409865686]
This paper proposes a deep learning (DL) model for automatic sleep stage classification based on single-channel EEG data.
The model was designed to run on energy and memory-constrained devices for real-time operation with local processing.
We tested a reduced-sized version of the proposed model on a low-cost Arduino Nano 33 BLE board and it was fully functional and accurate.
arXiv Detail & Related papers (2022-11-20T16:22:30Z) - Performance and utility trade-off in interpretable sleep staging [0.0]
We explore interpretable methods for a clinical decision support system, sleep staging, based on physiological signals such as EEG, EOG, and EMG.
A proposed framework, NormIntSleep, shows that by representing deep learning embeddings using normalized features, great performance can be obtained across different datasets.
arXiv Detail & Related papers (2022-11-07T03:27:01Z) - Slimmable Networks for Contrastive Self-supervised Learning [69.9454691873866]
Self-supervised learning makes significant progress in pre-training large models, but struggles with small models.
We introduce another one-stage solution to obtain pre-trained small models without the need for extra teachers.
A slimmable network consists of a full network and several weight-sharing sub-networks, which can be pre-trained once to obtain various networks.
arXiv Detail & Related papers (2022-09-30T15:15:05Z) - Incremental Online Learning Algorithms Comparison for Gesture and Visual
Smart Sensors [68.8204255655161]
This paper compares four state-of-the-art algorithms in two real applications: gesture recognition based on accelerometer data and image classification.
Our results confirm these systems' reliability and the feasibility of deploying them in tiny-memory MCUs.
arXiv Detail & Related papers (2022-09-01T17:05:20Z) - Which priors matter? Benchmarking models for learning latent dynamics [70.88999063639146]
Several methods have proposed to integrate priors from classical mechanics into machine learning models.
We take a sober look at the current capabilities of these models.
We find that the use of continuous and time-reversible dynamics benefits models of all classes.
arXiv Detail & Related papers (2021-11-09T23:48:21Z) - Sleep Staging Based on Serialized Dual Attention Network [0.0]
We propose a deep learning model SDAN based on raw EEG.
It serially combines the channel attention and spatial attention mechanisms to filter and highlight key information.
It achieves excellent results in the N1 sleep stage compared to other methods.
arXiv Detail & Related papers (2021-07-18T13:18:12Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.