Federated Learning of Medical Concepts Embedding using BEHRT
- URL: http://arxiv.org/abs/2305.13052v1
- Date: Mon, 22 May 2023 14:05:39 GMT
- Title: Federated Learning of Medical Concepts Embedding using BEHRT
- Authors: Ofir Ben Shoham, Nadav Rappoport
- Abstract summary: We propose a federated learning approach for learning medical concepts embedding.
Our approach is based on embedding model like BEHRT, a deep neural sequence model for EHR.
We compare the performance of a model trained with FL against a model trained on centralized data.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Electronic Health Records (EHR) data contains medical records such as
diagnoses, medications, procedures, and treatments of patients. This data is
often considered sensitive medical information. Therefore, the EHR data from
the medical centers often cannot be shared, making it difficult to create
prediction models using multi-center EHR data, which is essential for such
models' robustness and generalizability. Federated Learning (FL) is an
algorithmic approach that allows learning a shared model using data in multiple
locations without the need to store all data in a central place. An example of
a prediction model's task is to predict future diseases. More specifically, the
model needs to predict patient's next visit diagnoses, based on current and
previous clinical data. Such a prediction model can support care providers in
making clinical decisions and even provide preventive treatment. We propose a
federated learning approach for learning medical concepts embedding. This
pre-trained model can be used for fine-tuning for specific downstream tasks.
Our approach is based on an embedding model like BEHRT, a deep neural sequence
transduction model for EHR. We train using federated learning, both the Masked
Language Modeling (MLM) and the next visit downstream model. We demonstrate our
approach on the MIMIC-IV dataset. We compare the performance of a model trained
with FL against a model trained on centralized data. We find that our federated
learning approach reaches very close to the performance of a centralized model,
and it outperforms local models in terms of average precision. We also show
that pre-trained MLM improves the model's average precision performance in the
next visit prediction task, compared to an MLM model without pre-training. Our
code is available at https://github.com/nadavlab/FederatedBEHRT.
Related papers
- Extracting Training Data from Unconditional Diffusion Models [76.85077961718875]
diffusion probabilistic models (DPMs) are being employed as mainstream models for generative artificial intelligence (AI)
We aim to establish a theoretical understanding of memorization in DPMs with 1) a memorization metric for theoretical analysis, 2) an analysis of conditional memorization with informative and random labels, and 3) two better evaluation metrics for measuring memorization.
Based on the theoretical analysis, we propose a novel data extraction method called textbfSurrogate condItional Data Extraction (SIDE) that leverages a trained on generated data as a surrogate condition to extract training data directly from unconditional diffusion models.
arXiv Detail & Related papers (2024-06-18T16:20:12Z) - MedDiffusion: Boosting Health Risk Prediction via Diffusion-based Data
Augmentation [58.93221876843639]
This paper introduces a novel, end-to-end diffusion-based risk prediction model, named MedDiffusion.
It enhances risk prediction performance by creating synthetic patient data during training to enlarge sample space.
It discerns hidden relationships between patient visits using a step-wise attention mechanism, enabling the model to automatically retain the most vital information for generating high-quality data.
arXiv Detail & Related papers (2023-10-04T01:36:30Z) - EHRSHOT: An EHR Benchmark for Few-Shot Evaluation of Foundation Models [6.506937003687058]
We publish a new dataset, EHRSHOT, which contains deidentified structured data from the electronic health records (EHRs) of 6,739 patients from Stanford Medicine.
Second, we publish the weights of CLMBR-T-base, a 141M parameter clinical foundation model pretrained on the structured EHR data of 2.57M patients.
Third, we define 15 few-shot clinical prediction tasks, enabling evaluation of foundation models on benefits such as sample efficiency and task adaptation.
arXiv Detail & Related papers (2023-07-05T05:24:59Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Unsupervised Pre-Training on Patient Population Graphs for Patient-Level
Predictions [48.02011627390706]
Pre-training has shown success in different areas of machine learning, such as Computer Vision (CV), Natural Language Processing (NLP) and medical imaging.
In this paper, we apply unsupervised pre-training to heterogeneous, multi-modal EHR data for patient outcome prediction.
We find that our proposed graph based pre-training method helps in modeling the data at a population level.
arXiv Detail & Related papers (2022-03-23T17:59:45Z) - Multi-task fusion for improving mammography screening data
classification [3.7683182861690843]
We propose a pipeline approach, where we first train a set of individual, task-specific models.
We then investigate the fusion thereof, which is in contrast to the standard model ensembling strategy.
Our fusion approaches improve AUC scores significantly by up to 0.04 compared to standard model ensembling.
arXiv Detail & Related papers (2021-12-01T13:56:27Z) - Pre-training transformer-based framework on large-scale pediatric claims
data for downstream population-specific tasks [3.1580072841682734]
This study presents the Claim Pre-Training (Claim-PT) framework, a generic pre-training model that first trains on the entire pediatric claims dataset.
The effective knowledge transfer is completed through the task-aware fine-tuning stage.
We conducted experiments on a real-world claims dataset with more than one million patient records.
arXiv Detail & Related papers (2021-06-24T15:25:41Z) - EventScore: An Automated Real-time Early Warning Score for Clinical
Events [3.3039612529376625]
We build an interpretable model for the early prediction of various adverse clinical events indicative of clinical deterioration.
The model is evaluated on two datasets and four clinical events.
Our model can be entirely automated without requiring any manually recorded features.
arXiv Detail & Related papers (2021-02-11T11:55:08Z) - Self-Training with Improved Regularization for Sample-Efficient Chest
X-Ray Classification [80.00316465793702]
We present a deep learning framework that enables robust modeling in challenging scenarios.
Our results show that using 85% lesser labeled data, we can build predictive models that match the performance of classifiers trained in a large-scale data setting.
arXiv Detail & Related papers (2020-05-03T02:36:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.