Sequential Multi-Dimensional Self-Supervised Learning for Clinical Time
Series
- URL: http://arxiv.org/abs/2307.10923v1
- Date: Thu, 20 Jul 2023 14:49:58 GMT
- Title: Sequential Multi-Dimensional Self-Supervised Learning for Clinical Time
Series
- Authors: Aniruddh Raghu, Payal Chandak, Ridwan Alam, John Guttag, Collin M.
Stultz
- Abstract summary: We propose a new self-supervised learning method for clinical time series data.
Our method is agnostic to the specific form of loss function used at each level.
We evaluate our method on two real-world clinical datasets.
- Score: 3.635056427544418
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised learning (SSL) for clinical time series data has received
significant attention in recent literature, since these data are highly rich
and provide important information about a patient's physiological state.
However, most existing SSL methods for clinical time series are limited in that
they are designed for unimodal time series, such as a sequence of structured
features (e.g., lab values and vitals signs) or an individual high-dimensional
physiological signal (e.g., an electrocardiogram). These existing methods
cannot be readily extended to model time series that exhibit multimodality,
with structured features and high-dimensional data being recorded at each
timestep in the sequence. In this work, we address this gap and propose a new
SSL method -- Sequential Multi-Dimensional SSL -- where a SSL loss is applied
both at the level of the entire sequence and at the level of the individual
high-dimensional data points in the sequence in order to better capture
information at both scales. Our strategy is agnostic to the specific form of
loss function used at each level -- it can be contrastive, as in SimCLR, or
non-contrastive, as in VICReg. We evaluate our method on two real-world
clinical datasets, where the time series contains sequences of (1)
high-frequency electrocardiograms and (2) structured data from lab values and
vitals signs. Our experimental results indicate that pre-training with our
method and then fine-tuning on downstream tasks improves performance over
baselines on both datasets, and in several settings, can lead to improvements
across different self-supervised loss functions.
Related papers
- FedCVD: The First Real-World Federated Learning Benchmark on Cardiovascular Disease Data [52.55123685248105]
Cardiovascular diseases (CVDs) are currently the leading cause of death worldwide, highlighting the critical need for early diagnosis and treatment.
Machine learning (ML) methods can help diagnose CVDs early, but their performance relies on access to substantial data with high quality.
This paper presents the first real-world FL benchmark for cardiovascular disease detection, named FedCVD.
arXiv Detail & Related papers (2024-10-28T02:24:01Z) - MedTsLLM: Leveraging LLMs for Multimodal Medical Time Series Analysis [6.30440420617113]
We introduce MedTsLLM, a general multimodal large language model (LLM) framework that integrates time series data and rich contextual information in the form of text to analyze physiological signals.
We perform three tasks with clinical relevance: semantic segmentation, boundary detection, and anomaly detection in time series.
Our model outperforms state-of-the-art baselines, including deep learning models, other LLMs, and clinical methods across multiple medical domains.
arXiv Detail & Related papers (2024-08-14T18:57:05Z) - Temporal Cross-Attention for Dynamic Embedding and Tokenization of Multimodal Electronic Health Records [1.6609516435725236]
We introduce a dynamic embedding and tokenization framework for precise representation of multimodal clinical time series.
Our framework outperformed baseline approaches on the task of predicting the occurrence of nine postoperative complications.
arXiv Detail & Related papers (2024-03-06T19:46:44Z) - On the Importance of Step-wise Embeddings for Heterogeneous Clinical
Time-Series [1.3285222309805063]
Recent advances in deep learning for sequence modeling have not fully transferred to tasks handling time-series from electronic health records.
In particular, in problems related to the Intensive Care Unit (ICU), the state-of-the-art remains to tackle sequence classification in a tabular manner with tree-based methods.
arXiv Detail & Related papers (2023-11-15T12:18:15Z) - ICU Mortality Prediction Using Long Short-Term Memory Networks [0.0]
We implement an automatic data-driven system, which analyzes large amounts of temporal data derived from Electronic Health Records (EHRs)
We extract high-level information so as to predict in-hospital mortality and Length of Stay (LOS) early.
Experiments highlight the efficiency of LSTM model with rigorous time-series measurements for building real-world prediction engines.
arXiv Detail & Related papers (2023-08-18T09:44:47Z) - Self-Supervised Learning for Time Series Analysis: Taxonomy, Progress, and Prospects [84.6945070729684]
Self-supervised learning (SSL) has recently achieved impressive performance on various time series tasks.
This article reviews current state-of-the-art SSL methods for time series data.
arXiv Detail & Related papers (2023-06-16T18:23:10Z) - Multimodality Multi-Lead ECG Arrhythmia Classification using
Self-Supervised Learning [5.675787521359948]
We propose SSL-based multimodality ECG classification.
Our proposed network follows SSL learning paradigm and consists of two modules corresponding to pre-stream task, and down-stream task.
To evaluate the effectiveness of our approach, ten-fold cross validation on the 12-lead PhysioNet 2020 dataset has been conducted.
arXiv Detail & Related papers (2022-09-30T18:45:34Z) - LifeLonger: A Benchmark for Continual Disease Classification [59.13735398630546]
We introduce LifeLonger, a benchmark for continual disease classification on the MedMNIST collection.
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch.
Cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
arXiv Detail & Related papers (2022-04-12T12:25:05Z) - Federated Cycling (FedCy): Semi-supervised Federated Learning of
Surgical Phases [57.90226879210227]
FedCy is a semi-supervised learning (FSSL) method that combines FL and self-supervised learning to exploit a decentralized dataset of both labeled and unlabeled videos.
We demonstrate significant performance gains over state-of-the-art FSSL methods on the task of automatic recognition of surgical phases.
arXiv Detail & Related papers (2022-03-14T17:44:53Z) - Learning summary features of time series for likelihood free inference [93.08098361687722]
We present a data-driven strategy for automatically learning summary features from time series data.
Our results indicate that learning summary features from data can compete and even outperform LFI methods based on hand-crafted values.
arXiv Detail & Related papers (2020-12-04T19:21:37Z) - Uncovering the structure of clinical EEG signals with self-supervised
learning [64.4754948595556]
Supervised learning paradigms are often limited by the amount of labeled data that is available.
This phenomenon is particularly problematic in clinically-relevant data, such as electroencephalography (EEG)
By extracting information from unlabeled data, it might be possible to reach competitive performance with deep neural networks.
arXiv Detail & Related papers (2020-07-31T14:34:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.