Multi-Task Prediction of Clinical Outcomes in the Intensive Care Unit
using Flexible Multimodal Transformers
- URL: http://arxiv.org/abs/2111.05431v1
- Date: Tue, 9 Nov 2021 21:46:11 GMT
- Title: Multi-Task Prediction of Clinical Outcomes in the Intensive Care Unit
using Flexible Multimodal Transformers
- Authors: Benjamin Shickel, Patrick J. Tighe, Azra Bihorac, Parisa Rashidi
- Abstract summary: We propose a flexible Transformer-based EHR embedding pipeline and predictive model framework.
We showcase the feasibility of our flexible design in a case study in the intensive care unit.
- Score: 4.836546574465437
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent deep learning research based on Transformer model architectures has
demonstrated state-of-the-art performance across a variety of domains and
tasks, mostly within the computer vision and natural language processing
domains. While some recent studies have implemented Transformers for clinical
tasks using electronic health records data, they are limited in scope,
flexibility, and comprehensiveness. In this study, we propose a flexible
Transformer-based EHR embedding pipeline and predictive model framework that
introduces several novel modifications of existing workflows that capitalize on
data attributes unique to the healthcare domain. We showcase the feasibility of
our flexible design in a case study in the intensive care unit, where our
models accurately predict seven clinical outcomes pertaining to readmission and
patient mortality over multiple future time horizons.
Related papers
- MEDFuse: Multimodal EHR Data Fusion with Masked Lab-Test Modeling and Large Language Models [11.798375238713488]
MEDFuse is a framework that integrates structured and unstructured medical data.
It achieves over 90% F1 score in the 10-disease multi-label classification task.
arXiv Detail & Related papers (2024-07-17T04:17:09Z) - CViT: Continuous Vision Transformer for Operator Learning [24.1795082775376]
Continuous Vision Transformer (CViT) is a novel neural operator architecture that leverages advances in computer vision to address challenges in learning complex physical systems.
CViT combines a vision transformer encoder, a novel grid-based coordinate embedding, and a query-wise cross-attention mechanism to effectively capture multi-scale dependencies.
We demonstrate CViT's effectiveness across a diverse range of partial differential equation (PDE) systems, including fluid dynamics, climate modeling, and reaction-diffusion processes.
arXiv Detail & Related papers (2024-05-22T21:13:23Z) - Towards a clinically accessible radiology foundation model: open-access and lightweight, with automated evaluation [113.5002649181103]
Training open-source small multimodal models (SMMs) to bridge competency gaps for unmet clinical needs in radiology.
For training, we assemble a large dataset of over 697 thousand radiology image-text pairs.
For evaluation, we propose CheXprompt, a GPT-4-based metric for factuality evaluation, and demonstrate its parity with expert evaluation.
The inference of LlaVA-Rad is fast and can be performed on a single V100 GPU in private settings, offering a promising state-of-the-art tool for real-world clinical applications.
arXiv Detail & Related papers (2024-03-12T18:12:02Z) - Multi-modal Deep Learning [0.0]
The study refines clinical data processing through Compact Convolutional Transformer (CCT), Patch Up, and the innovative CamCenterLoss technique.
The proposed methodology demonstrates improved prediction accuracy and at tentiveness to critically ill patients compared to Guo JingYuan's ResNet and StageNet approaches.
arXiv Detail & Related papers (2024-03-06T00:36:05Z) - A Comprehensive Survey on Applications of Transformers for Deep Learning
Tasks [60.38369406877899]
Transformer is a deep neural network that employs a self-attention mechanism to comprehend the contextual relationships within sequential data.
transformer models excel in handling long dependencies between input sequence elements and enable parallel processing.
Our survey encompasses the identification of the top five application domains for transformer-based models.
arXiv Detail & Related papers (2023-06-11T23:13:51Z) - A Transformer-based representation-learning model with unified
processing of multimodal input for clinical diagnostics [63.106382317917344]
We report a Transformer-based representation-learning model as a clinical diagnostic aid that processes multimodal input in a unified manner.
The unified model outperformed an image-only model and non-unified multimodal diagnosis models in the identification of pulmonary diseases.
arXiv Detail & Related papers (2023-06-01T16:23:47Z) - MVMTnet: A Multi-variate Multi-modal Transformer for Multi-class
Classification of Cardiac Irregularities Using ECG Waveforms and Clinical
Notes [4.648677931378919]
Deep learning can be used to optimize diagnosis and patient monitoring for clinical-based applications.
For cardiovascular disease, one such condition where the rising number of patients increasingly outweighs the availability of medical resources in different parts of the world, a core challenge is the automated classification of various cardiac abnormalities.
The proposed novel multi-modal Transformer architecture would be able to accurately perform this task while demonstrating the cross-domain effectiveness of Transformers.
arXiv Detail & Related papers (2023-02-21T21:38:41Z) - An Experimental Evaluation of Transformer-based Language Models in the
Biomedical Domain [0.984441002699829]
This paper summarizes experiments conducted in replicating BioBERT and further pre-training and fine-tuning in the biomedical domain.
We also investigate the effectiveness of domain-specific and domain-agnostic pre-trained models across downstream biomedical NLP tasks.
arXiv Detail & Related papers (2020-12-31T03:09:38Z) - Towards an Automatic Analysis of CHO-K1 Suspension Growth in
Microfluidic Single-cell Cultivation [63.94623495501023]
We propose a novel Machine Learning architecture, which allows us to infuse a neural deep network with human-powered abstraction on the level of data.
Specifically, we train a generative model simultaneously on natural and synthetic data, so that it learns a shared representation, from which a target variable, such as the cell count, can be reliably estimated.
arXiv Detail & Related papers (2020-10-20T08:36:51Z) - Domain Shift in Computer Vision models for MRI data analysis: An
Overview [64.69150970967524]
Machine learning and computer vision methods are showing good performance in medical imagery analysis.
Yet only a few applications are now in clinical use.
Poor transferability of themodels to data from different sources or acquisition domains is one of the reasons for that.
arXiv Detail & Related papers (2020-10-14T16:34:21Z) - Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype
Prediction [55.94378672172967]
We focus on few-shot disease subtype prediction problem, identifying subgroups of similar patients.
We introduce meta learning techniques to develop a new model, which can extract the common experience or knowledge from interrelated clinical tasks.
Our new model is built upon a carefully designed meta-learner, called Prototypical Network, that is a simple yet effective meta learning machine for few-shot image classification.
arXiv Detail & Related papers (2020-09-02T02:50:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.