Data-Free Distillation Improves Efficiency and Privacy in Federated
Thorax Disease Analysis
- URL: http://arxiv.org/abs/2310.18346v2
- Date: Tue, 31 Oct 2023 09:13:18 GMT
- Title: Data-Free Distillation Improves Efficiency and Privacy in Federated
Thorax Disease Analysis
- Authors: Ming Li and Guang Yang
- Abstract summary: Thorax disease analysis in large-scale, multi-centre, and multi-scanner settings is often limited by strict privacy policies.
We introduce a data-free distillation-based FL approach FedKDF.
In FedKDF, the server employs a lightweight generator to aggregate knowledge from different clients without requiring access to their private data or a proxy dataset.
- Score: 11.412151951949102
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Thorax disease analysis in large-scale, multi-centre, and multi-scanner
settings is often limited by strict privacy policies. Federated learning (FL)
offers a potential solution, while traditional parameter-based FL can be
limited by issues such as high communication costs, data leakage, and
heterogeneity. Distillation-based FL can improve efficiency, but it relies on a
proxy dataset, which is often impractical in clinical practice. To address
these challenges, we introduce a data-free distillation-based FL approach
FedKDF. In FedKDF, the server employs a lightweight generator to aggregate
knowledge from different clients without requiring access to their private data
or a proxy dataset. FedKDF combines the predictors from clients into a single,
unified predictor, which is further optimized using the learned knowledge in
the lightweight generator. Our empirical experiments demonstrate that FedKDF
offers a robust solution for efficient, privacy-preserving federated thorax
disease analysis.
Related papers
- Subgraph Federated Learning via Spectral Methods [52.40322201034717]
FedLap is a novel framework that captures inter-node dependencies while ensuring privacy and scalability.<n>We provide a formal analysis of the privacy of FedLap, demonstrating that it preserves privacy.
arXiv Detail & Related papers (2025-10-29T16:22:32Z) - A Robust Pipeline for Differentially Private Federated Learning on Imbalanced Clinical Data using SMOTETomek and FedProx [0.0]
Federated Learning (FL) presents a groundbreaking approach for collaborative health research.<n>FL offers formal security guarantees when combined with Differential Privacy (DP)<n>An optimal operational region was identified on the privacy-utility frontier.
arXiv Detail & Related papers (2025-08-06T20:47:50Z) - Blockchain-Enabled Privacy-Preserving Second-Order Federated Edge Learning in Personalized Healthcare [1.859970493489417]
Federated learning (FL) has attracted increasing attention to security and privacy challenges in traditional cloud-centric machine learning models.<n>First-order FL approaches face several challenges in personalized model training due to heterogeneous non-independent and identically distributed (non-iid) data.<n>Recently, second-order FL approaches maintain the stability and consistency of non-iid datasets while improving personalized model training.
arXiv Detail & Related papers (2025-05-31T06:41:04Z) - Fine-Tuning Foundation Models with Federated Learning for Privacy Preserving Medical Time Series Forecasting [0.32985979395737786]
Federated Learning (FL) provides a decentralized machine learning approach, where multiple devices or servers collaboratively train a model without sharing their raw data.
In this paper, we fine-tune time series FMs with Electrocardiogram (ECG) and Impedance Cardiography (ICG) data using different FL techniques.
Our empirical results demonstrated that while FL can be effective for fine-tuning FMs on time series forecasting tasks, its benefits depend on the data distribution across clients.
arXiv Detail & Related papers (2025-02-13T20:01:15Z) - Privacy-preserving Federated Primal-dual Learning for Non-convex and Non-smooth Problems with Model Sparsification [51.04894019092156]
Federated learning (FL) has been recognized as a rapidly growing area, where the model is trained over clients under the FL orchestration (PS)
In this paper, we propose a novel primal sparification algorithm for and guarantee non-smooth FL problems.
Its unique insightful properties and its analyses are also presented.
arXiv Detail & Related papers (2023-10-30T14:15:47Z) - PS-FedGAN: An Efficient Federated Learning Framework Based on Partially
Shared Generative Adversarial Networks For Data Privacy [56.347786940414935]
Federated Learning (FL) has emerged as an effective learning paradigm for distributed computation.
This work proposes a novel FL framework that requires only partial GAN model sharing.
Named as PS-FedGAN, this new framework enhances the GAN releasing and training mechanism to address heterogeneous data distributions.
arXiv Detail & Related papers (2023-05-19T05:39:40Z) - Federated Learning with Uncertainty-Based Client Clustering for
Fleet-Wide Fault Diagnosis [5.684489493030555]
Federated learning has emerged as a promising solution to leverage datasets from multiple operators to train a decentralized asset fault diagnosis model.
We propose a novel clustering-based FL algorithm where clients are clustered for federating based on dataset similarity.
arXiv Detail & Related papers (2023-04-26T04:23:59Z) - FedSkip: Combatting Statistical Heterogeneity with Federated Skip
Aggregation [95.85026305874824]
We introduce a data-driven approach called FedSkip to improve the client optima by periodically skipping federated averaging and scattering local models to the cross devices.
We conduct extensive experiments on a range of datasets to demonstrate that FedSkip achieves much higher accuracy, better aggregation efficiency and competing communication efficiency.
arXiv Detail & Related papers (2022-12-14T13:57:01Z) - Federated Learning with Privacy-Preserving Ensemble Attention
Distillation [63.39442596910485]
Federated Learning (FL) is a machine learning paradigm where many local nodes collaboratively train a central model while keeping the training data decentralized.
We propose a privacy-preserving FL framework leveraging unlabeled public data for one-way offline knowledge distillation.
Our technique uses decentralized and heterogeneous local data like existing FL approaches, but more importantly, it significantly reduces the risk of privacy leakage.
arXiv Detail & Related papers (2022-10-16T06:44:46Z) - Fine-tuning Global Model via Data-Free Knowledge Distillation for
Non-IID Federated Learning [86.59588262014456]
Federated Learning (FL) is an emerging distributed learning paradigm under privacy constraint.
We propose a data-free knowledge distillation method to fine-tune the global model in the server (FedFTG)
Our FedFTG significantly outperforms the state-of-the-art (SOTA) FL algorithms and can serve as a strong plugin for enhancing FedAvg, FedProx, FedDyn, and SCAFFOLD.
arXiv Detail & Related papers (2022-03-17T11:18:17Z) - Improving Performance of Federated Learning based Medical Image Analysis
in Non-IID Settings using Image Augmentation [1.5469452301122177]
Federated Learning (FL) is a suitable solution for making use of sensitive data belonging to patients, people, companies, or industries that are obligatory to work under rigid privacy constraints.
FL mainly or partially supports data privacy and security issues and provides an alternative to model problems facilitating multiple edge devices or organizations to contribute a training of a global model using a number of local data without having them.
This paper introduces a novel method dynamically balancing the data distributions of clients by augmenting images to address the non-IID data problem of FL.
arXiv Detail & Related papers (2021-12-12T10:05:42Z) - Local Learning Matters: Rethinking Data Heterogeneity in Federated
Learning [61.488646649045215]
Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices)
arXiv Detail & Related papers (2021-11-28T19:03:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.