Performance and utility trade-off in interpretable sleep staging
- URL: http://arxiv.org/abs/2211.03282v1
- Date: Mon, 7 Nov 2022 03:27:01 GMT
- Title: Performance and utility trade-off in interpretable sleep staging
- Authors: Irfan Al-Hussaini, Cassie S. Mitchell
- Abstract summary: We explore interpretable methods for a clinical decision support system, sleep staging, based on physiological signals such as EEG, EOG, and EMG.
A proposed framework, NormIntSleep, shows that by representing deep learning embeddings using normalized features, great performance can be obtained across different datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances in deep learning have led to the development of models
approaching human level of accuracy. However, healthcare remains an area
lacking in widespread adoption. The safety-critical nature of healthcare
results in a natural reticence to put these black-box deep learning models into
practice. In this paper, we explore interpretable methods for a clinical
decision support system, sleep staging, based on physiological signals such as
EEG, EOG, and EMG. A recent work has shown sleep staging using simple models
and an exhaustive set of features can perform nearly as well as deep learning
approaches but only for certain datasets. Moreover, the utility of these
features from a clinical standpoint is unclear. On the other hand, the proposed
framework, NormIntSleep shows that by representing deep learning embeddings
using normalized features, great performance can be obtained across different
datasets. NormIntSleep performs 4.5% better than the exhaustive feature-based
approach and 1.5% better than other representation learning approaches. An
empirical comparison between the utility of the interpretations of these models
highlights the improved alignment with clinical expectations when performance
is traded-off slightly.
Related papers
- On the Out of Distribution Robustness of Foundation Models in Medical
Image Segmentation [47.95611203419802]
Foundations for vision and language, pre-trained on extensive sets of natural image and text data, have emerged as a promising approach.
We compare the generalization performance to unseen domains of various pre-trained models after being fine-tuned on the same in-distribution dataset.
We further developed a new Bayesian uncertainty estimation for frozen models and used them as an indicator to characterize the model's performance on out-of-distribution data.
arXiv Detail & Related papers (2023-11-18T14:52:10Z) - Transparency in Sleep Staging: Deep Learning Method for EEG Sleep Stage
Classification with Model Interpretability [5.747465732334616]
This study presents an end-to-end deep learning (DL) model which integrates squeeze and excitation blocks within the residual network to extract features and stacked Bi-LSTM to understand complex temporal dependencies.
A distinctive aspect of this study is the adaptation of GradCam for sleep staging, marking the first instance of an explainable DL model in this domain with alignment of its decision-making with sleep expert's insights.
arXiv Detail & Related papers (2023-09-10T17:56:03Z) - TREEMENT: Interpretable Patient-Trial Matching via Personalized Dynamic
Tree-Based Memory Network [54.332862955411656]
Clinical trials are critical for drug development but often suffer from expensive and inefficient patient recruitment.
In recent years, machine learning models have been proposed for speeding up patient recruitment via automatically matching patients with clinical trials.
We introduce a dynamic tree-based memory network model named TREEMENT to provide accurate and interpretable patient trial matching.
arXiv Detail & Related papers (2023-07-19T12:35:09Z) - Rethinking Semi-Supervised Medical Image Segmentation: A
Variance-Reduction Perspective [51.70661197256033]
We propose ARCO, a semi-supervised contrastive learning framework with stratified group theory for medical image segmentation.
We first propose building ARCO through the concept of variance-reduced estimation and show that certain variance-reduction techniques are particularly beneficial in pixel/voxel-level segmentation tasks.
We experimentally validate our approaches on eight benchmarks, i.e., five 2D/3D medical and three semantic segmentation datasets, with different label settings.
arXiv Detail & Related papers (2023-02-03T13:50:25Z) - Do Not Sleep on Linear Models: Simple and Interpretable Techniques
Outperform Deep Learning for Sleep Scoring [1.6339105551302067]
We argue that most deep learning solutions for sleep scoring are limited in their real-world applicability as they are hard to train, deploy, and reproduce.
In this work, we revisit the problem of sleep stage classification using classical machine learning.
Results show that state-of-the-art performance can be achieved with a conventional machine learning pipeline.
arXiv Detail & Related papers (2022-07-15T21:03:11Z) - Benchmarking Heterogeneous Treatment Effect Models through the Lens of
Interpretability [82.29775890542967]
Estimating personalized effects of treatments is a complex, yet pervasive problem.
Recent developments in the machine learning literature on heterogeneous treatment effect estimation gave rise to many sophisticated, but opaque, tools.
We use post-hoc feature importance methods to identify features that influence the model's predictions.
arXiv Detail & Related papers (2022-06-16T17:59:05Z) - LifeLonger: A Benchmark for Continual Disease Classification [59.13735398630546]
We introduce LifeLonger, a benchmark for continual disease classification on the MedMNIST collection.
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch.
Cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
arXiv Detail & Related papers (2022-04-12T12:25:05Z) - Enhancing Clinical Information Extraction with Transferred Contextual
Embeddings [9.143551270841858]
Bidirectional Representations from Transformers (BERT) model has achieved the state-of-the-art performance for many natural language processing (NLP) tasks.
We show that BERT based pre-training models can be transferred to health-related documents under mild conditions.
arXiv Detail & Related papers (2021-09-15T12:22:57Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - Learning Realistic Patterns from Unrealistic Stimuli: Generalization and
Data Anonymization [0.5091527753265949]
This work investigates a simple yet unconventional approach for anonymized data synthesis to enable third parties to benefit from such private data.
We use sleep monitoring data from both an open and a large closed clinical study and evaluate whether (1) end-users can create and successfully use customized classification models for sleep apnea detection, and (2) the identity of participants in the study is protected.
arXiv Detail & Related papers (2020-09-21T16:31:21Z) - MetaSleepLearner: A Pilot Study on Fast Adaptation of Bio-signals-Based
Sleep Stage Classifier to New Individual Subject Using Meta-Learning [15.451212330924447]
We propose the transfer learning framework, entitled MetaSleepLearner, based on Model Agnostic Meta-Learning (MAML)
In comparison to the conventional approach, MetaSleepLearner achieved a range of 5.4% to 17.7% improvement with statistical difference in the mean of both approaches.
This is the first work that investigated a non-conventional pre-training method, MAML, resulting in a possibility for human-machine collaboration in sleep stage classification.
arXiv Detail & Related papers (2020-04-08T16:31:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.