Med-BERT: pre-trained contextualized embeddings on large-scale
structured electronic health records for disease prediction
- URL: http://arxiv.org/abs/2005.12833v1
- Date: Fri, 22 May 2020 05:07:17 GMT
- Title: Med-BERT: pre-trained contextualized embeddings on large-scale
structured electronic health records for disease prediction
- Authors: Laila Rasmy, Yang Xiang, Ziqian Xie, Cui Tao and Degui Zhi
- Abstract summary: We propose Med-BERT, which adapts the BERT framework for pre-training contextualized embedding models on structured diagnosis data from 28,490,650 patients EHR dataset.
Med-BERT substantially improves prediction accuracy, boosting the area under receiver operating characteristics curve (AUC) by 2.02-7.12%.
In particular, pre-trained Med-BERT substantially improves the performance of tasks with very small fine-tuning training sets (300-500 samples) boosting the AUC by more than 20% or equivalent to the AUC of 10 times larger training set.
- Score: 12.669003066030697
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning (DL) based predictive models from electronic health records
(EHR) deliver impressive performance in many clinical tasks. Large training
cohorts, however, are often required to achieve high accuracy, hindering the
adoption of DL-based models in scenarios with limited training data size.
Recently, bidirectional encoder representations from transformers (BERT) and
related models have achieved tremendous successes in the natural language
processing domain. The pre-training of BERT on a very large training corpus
generates contextualized embeddings that can boost the performance of models
trained on smaller datasets. We propose Med-BERT, which adapts the BERT
framework for pre-training contextualized embedding models on structured
diagnosis data from 28,490,650 patients EHR dataset. Fine-tuning experiments
are conducted on two disease-prediction tasks: (1) prediction of heart failure
in patients with diabetes and (2) prediction of pancreatic cancer from two
clinical databases. Med-BERT substantially improves prediction accuracy,
boosting the area under receiver operating characteristics curve (AUC) by
2.02-7.12%. In particular, pre-trained Med-BERT substantially improves the
performance of tasks with very small fine-tuning training sets (300-500
samples) boosting the AUC by more than 20% or equivalent to the AUC of 10 times
larger training set. We believe that Med-BERT will benefit disease-prediction
studies with small local training datasets, reduce data collection expenses,
and accelerate the pace of artificial intelligence aided healthcare.
Related papers
- BISeizuRe: BERT-Inspired Seizure Data Representation to Improve Epilepsy Monitoring [13.35453284825286]
This study presents a novel approach for EEG-based seizure detection leveraging a BERT-based model.
The model, BENDR, undergoes a two-phase training process, pre-training and fine-tuning.
The optimized model demonstrates substantial performance enhancements, achieving as low as 0.23 FP/h, 2.5$times$ lower than the baseline model, with a lower but still acceptable sensitivity rate.
arXiv Detail & Related papers (2024-06-27T14:09:10Z) - Clinical Deterioration Prediction in Brazilian Hospitals Based on
Artificial Neural Networks and Tree Decision Models [56.93322937189087]
An extremely boosted neural network (XBNet) is used to predict clinical deterioration (CD)
The XGBoost model obtained the best results in predicting CD among Brazilian hospitals' data.
arXiv Detail & Related papers (2022-12-17T23:29:14Z) - Textual Data Augmentation for Patient Outcomes Prediction [67.72545656557858]
We propose a novel data augmentation method to generate artificial clinical notes in patients' Electronic Health Records.
We fine-tune the generative language model GPT-2 to synthesize labeled text with the original training data.
We evaluate our method on the most common patient outcome, i.e., the 30-day readmission rate.
arXiv Detail & Related papers (2022-11-13T01:07:23Z) - A Comprehensive Benchmark for COVID-19 Predictive Modeling Using
Electronic Health Records in Intensive Care [15.64030213048907]
We propose two clinical prediction tasks, Outcome-specific length-of-stay prediction and Early mortality prediction for COVID-19 patients in intensive care units.
The two tasks are adapted from the naive length-of-stay and mortality prediction tasks to accommodate the clinical practice for COVID-19 patients.
We propose fair, detailed, open-source data-preprocessing pipelines and evaluate 17 state-of-the-art predictive models on two tasks.
arXiv Detail & Related papers (2022-09-16T09:09:15Z) - Density-Aware Personalized Training for Risk Prediction in Imbalanced
Medical Data [89.79617468457393]
Training models with imbalance rate (class density discrepancy) may lead to suboptimal prediction.
We propose a framework for training models for this imbalance issue.
We demonstrate our model's improved performance in real-world medical datasets.
arXiv Detail & Related papers (2022-07-23T00:39:53Z) - Unsupervised Pre-Training on Patient Population Graphs for Patient-Level
Predictions [48.02011627390706]
Pre-training has shown success in different areas of machine learning, such as Computer Vision (CV), Natural Language Processing (NLP) and medical imaging.
In this paper, we apply unsupervised pre-training to heterogeneous, multi-modal EHR data for patient outcome prediction.
We find that our proposed graph based pre-training method helps in modeling the data at a population level.
arXiv Detail & Related papers (2022-03-23T17:59:45Z) - CEHR-BERT: Incorporating temporal information from structured EHR data
to improve prediction tasks [0.0]
We develop a new BERT adaptation, CEHR-BERT, to incorporate temporal information using a hybrid approach.
CEHR-BERT was trained on a subset of Columbia University Irving Medical Center-York Presbyterian Hospital's clinical data.
arXiv Detail & Related papers (2021-11-10T16:53:32Z) - Pre-training transformer-based framework on large-scale pediatric claims
data for downstream population-specific tasks [3.1580072841682734]
This study presents the Claim Pre-Training (Claim-PT) framework, a generic pre-training model that first trains on the entire pediatric claims dataset.
The effective knowledge transfer is completed through the task-aware fine-tuning stage.
We conducted experiments on a real-world claims dataset with more than one million patient records.
arXiv Detail & Related papers (2021-06-24T15:25:41Z) - Predicting Clinical Diagnosis from Patients Electronic Health Records
Using BERT-based Neural Networks [62.9447303059342]
We show the importance of this problem in medical community.
We present a modification of Bidirectional Representations from Transformers (BERT) model for classification sequence.
We use a large-scale Russian EHR dataset consisting of about 4 million unique patient visits.
arXiv Detail & Related papers (2020-07-15T09:22:55Z) - Self-Training with Improved Regularization for Sample-Efficient Chest
X-Ray Classification [80.00316465793702]
We present a deep learning framework that enables robust modeling in challenging scenarios.
Our results show that using 85% lesser labeled data, we can build predictive models that match the performance of classifiers trained in a large-scale data setting.
arXiv Detail & Related papers (2020-05-03T02:36:00Z) - Language Models Are An Effective Patient Representation Learning
Technique For Electronic Health Record Data [7.260199064831896]
We show that patient representation schemes inspired from techniques in natural language processing can increase the accuracy of clinical prediction models.
Such patient representation schemes enable a 3.5% mean improvement in AUROC on five prediction tasks compared to standard baselines.
arXiv Detail & Related papers (2020-01-06T22:24:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.