FedNER: Privacy-preserving Medical Named Entity Recognition with
Federated Learning
- URL: http://arxiv.org/abs/2003.09288v2
- Date: Wed, 25 Mar 2020 06:37:06 GMT
- Title: FedNER: Privacy-preserving Medical Named Entity Recognition with
Federated Learning
- Authors: Suyu Ge, Fangzhao Wu, Chuhan Wu, Tao Qi, Yongfeng Huang, and Xing Xie
- Abstract summary: We propose a privacy-preserving medical NER method based on federated learning.
We decompose the medical NER model in each platform into a shared module and a private module.
The private module is used to capture the characteristics of the local data in each platform, and is updated using local labeled data.
- Score: 87.14713333892112
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical named entity recognition (NER) has wide applications in intelligent
healthcare. Sufficient labeled data is critical for training accurate medical
NER model. However, the labeled data in a single medical platform is usually
limited. Although labeled datasets may exist in many different medical
platforms, they cannot be directly shared since medical data is highly
privacy-sensitive. In this paper, we propose a privacy-preserving medical NER
method based on federated learning, which can leverage the labeled data in
different platforms to boost the training of medical NER model and remove the
need of exchanging raw data among different platforms. Since the labeled data
in different platforms usually has some differences in entity type and
annotation criteria, instead of constraining different platforms to share the
same model, we decompose the medical NER model in each platform into a shared
module and a private module. The private module is used to capture the
characteristics of the local data in each platform, and is updated using local
labeled data. The shared module is learned across different medical platform to
capture the shared NER knowledge. Its local gradients from different platforms
are aggregated to update the global shared module, which is further delivered
to each platform to update their local shared modules. Experiments on three
publicly available datasets validate the effectiveness of our method.
Related papers
- Universal Medical Imaging Model for Domain Generalization with Data Privacy [2.8727695958743364]
We propose a federated learning approach to transfer knowledge from multiple local models to a global model.
The primary objective is to train a global model capable of performing a wide variety of medical imaging tasks.
arXiv Detail & Related papers (2024-07-20T01:24:15Z) - Multi-Modal Dataset Creation for Federated Learning with DICOM Structured Reports [26.2463670182172]
Federated training is often hindered by heterogeneous datasets due to divergent data storage options, inconsistent naming schemes, varied annotation procedures, and disparities in label quality.
This is particularly evident in the emerging multi-modal learning paradigms, where dataset harmonization including a uniform data representation and filtering options are of paramount importance.
We developed an open platform for data integration and interactive filtering capabilities that simplifies the process of assembling multi-modal datasets.
arXiv Detail & Related papers (2024-07-12T07:34:10Z) - Feasibility of Federated Learning from Client Databases with Different Brain Diseases and MRI Modalities [5.629645463085369]
Models for brain lesions in MRI are typically developed for a specific disease and trained on data with a predefined set of MRI modalities.
This training paradigm prevents a model from using the advantages of learning from heterogeneous databases.
We demonstrate promising results by combining appropriate, simple, and practical modifications to the model and training strategy.
arXiv Detail & Related papers (2024-06-17T15:16:18Z) - UniCell: Universal Cell Nucleus Classification via Prompt Learning [76.11864242047074]
We propose a universal cell nucleus classification framework (UniCell)
It employs a novel prompt learning mechanism to uniformly predict the corresponding categories of pathological images from different dataset domains.
In particular, our framework adopts an end-to-end architecture for nuclei detection and classification, and utilizes flexible prediction heads for adapting various datasets.
arXiv Detail & Related papers (2024-02-20T11:50:27Z) - Rethinking Semi-Supervised Federated Learning: How to co-train
fully-labeled and fully-unlabeled client imaging data [6.322831694506287]
Isolated Federated Learning (IsoFed) is a learning scheme specifically designed for semi-supervised federated learning (SSFL)
We propose a novel learning scheme specifically designed for SSFL that circumvents the problem by avoiding simple averaging of supervised and semi-supervised models together.
In particular, our training approach consists of two parts - (a) isolated aggregation of labeled and unlabeled client models, and (b) local self-supervised pretraining of isolated global models in all clients.
arXiv Detail & Related papers (2023-10-28T20:41:41Z) - Learnable Weight Initialization for Volumetric Medical Image Segmentation [66.3030435676252]
We propose a learnable weight-based hybrid medical image segmentation approach.
Our approach is easy to integrate into any hybrid model and requires no external training data.
Experiments on multi-organ and lung cancer segmentation tasks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-06-15T17:55:05Z) - Federated Self-Supervised Contrastive Learning and Masked Autoencoder
for Dermatological Disease Diagnosis [15.20791611477636]
In dermatological disease diagnosis, the private data collected by mobile dermatology assistants exist on distributed mobile devices of patients.
We propose two federated self-supervised learning frameworks for dermatological disease diagnosis with limited labels.
Experiments on dermatological disease datasets show superior accuracy of the proposed frameworks over state-of-the-arts.
arXiv Detail & Related papers (2022-08-24T02:49:35Z) - Game of Privacy: Towards Better Federated Platform Collaboration under
Privacy Restriction [95.12382372267724]
Vertical federated learning (VFL) aims to train models from cross-silo data with different feature spaces stored on different platforms.
Due to the intrinsic privacy risks of federated learning, the total amount of involved data may be constrained.
We propose to incent different platforms through a reciprocal collaboration, where all platforms can exploit multi-platform information in the VFL framework to benefit their own tasks.
arXiv Detail & Related papers (2022-02-10T16:45:40Z) - Practical Challenges in Differentially-Private Federated Survival
Analysis of Medical Data [57.19441629270029]
In this paper, we take advantage of the inherent properties of neural networks to federate the process of training of survival analysis models.
In the realistic setting of small medical datasets and only a few data centers, this noise makes it harder for the models to converge.
We propose DPFed-post which adds a post-processing stage to the private federated learning scheme.
arXiv Detail & Related papers (2022-02-08T10:03:24Z) - Cross-Modal Information Maximization for Medical Imaging: CMIM [62.28852442561818]
In hospitals, data are siloed to specific information systems that make the same information available under different modalities.
This offers unique opportunities to obtain and use at train-time those multiple views of the same information that might not always be available at test-time.
We propose an innovative framework that makes the most of available data by learning good representations of a multi-modal input that are resilient to modality dropping at test-time.
arXiv Detail & Related papers (2020-10-20T20:05:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.