Unifying Distillation with Personalization in Federated Learning
- URL: http://arxiv.org/abs/2105.15191v1
- Date: Mon, 31 May 2021 17:54:29 GMT
- Title: Unifying Distillation with Personalization in Federated Learning
- Authors: Siddharth Divi, Habiba Farrukh, Berkay Celik
- Abstract summary: Federated learning (FL) is a decentralized privacy-preserving learning technique in which clients learn a joint collaborative model through a central aggregator without sharing their data.
In this setting, all clients learn a single common predictor (FedAvg), which does not generalize well on each client's local data due to the statistical data heterogeneity among clients.
In this paper, we address this problem with PersFL, a two-stage personalized learning algorithm.
In the first stage, PersFL finds the optimal teacher model of each client during the FL training phase. In the second stage, PersFL distills the useful knowledge from
- Score: 1.8262547855491458
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning (FL) is a decentralized privacy-preserving learning
technique in which clients learn a joint collaborative model through a central
aggregator without sharing their data. In this setting, all clients learn a
single common predictor (FedAvg), which does not generalize well on each
client's local data due to the statistical data heterogeneity among clients. In
this paper, we address this problem with PersFL, a discrete two-stage
personalized learning algorithm. In the first stage, PersFL finds the optimal
teacher model of each client during the FL training phase. In the second stage,
PersFL distills the useful knowledge from optimal teachers into each user's
local model. The teacher model provides each client with some rich, high-level
representation that a client can easily adapt to its local model, which
overcomes the statistical heterogeneity present at different clients. We
evaluate PersFL on CIFAR-10 and MNIST datasets using three data-splitting
strategies to control the diversity between clients' data distributions. We
empirically show that PersFL outperforms FedAvg and three state-of-the-art
personalization methods, pFedMe, Per-FedAvg, and FedPer on majority data-splits
with minimal communication cost. Further, we study the performance of PersFL on
different distillation objectives, how this performance is affected by the
equitable notion of fairness among clients, and the number of required
communication rounds. PersFL code is available at https://tinyurl.com/hdh5zhxs
for public use and validation.
Related papers
- FedReMa: Improving Personalized Federated Learning via Leveraging the Most Relevant Clients [13.98392319567057]
Federated Learning (FL) is a distributed machine learning paradigm that achieves a globally robust model through decentralized computation and periodic model synthesis.
Despite their wide adoption, existing FL and PFL works have yet to comprehensively address the class-imbalance issue.
We propose FedReMa, an efficient PFL algorithm that can tackle class-imbalance by utilizing an adaptive inter-client co-learning approach.
arXiv Detail & Related papers (2024-11-04T05:44:28Z) - An Aggregation-Free Federated Learning for Tackling Data Heterogeneity [50.44021981013037]
Federated Learning (FL) relies on the effectiveness of utilizing knowledge from distributed datasets.
Traditional FL methods adopt an aggregate-then-adapt framework, where clients update local models based on a global model aggregated by the server from the previous training round.
We introduce FedAF, a novel aggregation-free FL algorithm.
arXiv Detail & Related papers (2024-04-29T05:55:23Z) - PFL-GAN: When Client Heterogeneity Meets Generative Models in
Personalized Federated Learning [55.930403371398114]
We propose a novel generative adversarial network (GAN) sharing and aggregation strategy for personalized learning (PFL)
PFL-GAN addresses the client heterogeneity in different scenarios. More specially, we first learn the similarity among clients and then develop an weighted collaborative data aggregation.
The empirical results through the rigorous experimentation on several well-known datasets demonstrate the effectiveness of PFL-GAN.
arXiv Detail & Related papers (2023-08-23T22:38:35Z) - FedSampling: A Better Sampling Strategy for Federated Learning [81.85411484302952]
Federated learning (FL) is an important technique for learning models from decentralized data in a privacy-preserving way.
Existing FL methods usually uniformly sample clients for local model learning in each round.
We propose a novel data uniform sampling strategy for federated learning (FedSampling)
arXiv Detail & Related papers (2023-06-25T13:38:51Z) - Personalized Federated Learning on Long-Tailed Data via Adversarial
Feature Augmentation [24.679535905451758]
PFL aims to learn personalized models for each client based on the knowledge across all clients in a privacy-preserving manner.
Existing PFL methods assume that the underlying global data across all clients are uniformly distributed without considering the long-tail distribution.
We propose Federated Learning with Adversarial Feature Augmentation (FedAFA) to address this joint problem in PFL.
arXiv Detail & Related papers (2023-03-27T13:00:20Z) - Visual Prompt Based Personalized Federated Learning [83.04104655903846]
We propose a novel PFL framework for image classification tasks, dubbed pFedPT, that leverages personalized visual prompts to implicitly represent local data distribution information of clients.
Experiments on the CIFAR10 and CIFAR100 datasets show that pFedPT outperforms several state-of-the-art (SOTA) PFL algorithms by a large margin in various settings.
arXiv Detail & Related papers (2023-03-15T15:02:15Z) - New Metrics to Evaluate the Performance and Fairness of Personalized
Federated Learning [5.500172106704342]
In Federated Learning (FL), the clients learn a single global model (FedAvg) through a central aggregator.
In this setting, the non-IID distribution of the data across clients restricts the global FL model from delivering good performance on the local data of each client.
Personalized FL aims to address this problem by finding a personalized model for each client.
arXiv Detail & Related papers (2021-07-28T05:30:17Z) - FedProf: Optimizing Federated Learning with Dynamic Data Profiling [9.74942069718191]
Federated Learning (FL) has shown great potential as a privacy-preserving solution to learning from decentralized data.
A large proportion of the clients are probably in possession of only low-quality data that are biased, noisy or even irrelevant.
We propose a novel approach to optimizing FL under such circumstances without breaching data privacy.
arXiv Detail & Related papers (2021-02-02T20:10:14Z) - Personalized Federated Learning with First Order Model Optimization [76.81546598985159]
We propose an alternative to federated learning, where each client federates with other relevant clients to obtain a stronger model per client-specific objectives.
We do not assume knowledge of underlying data distributions or client similarities, and allow each client to optimize for arbitrary target distributions of interest.
Our method outperforms existing alternatives, while also enabling new features for personalized FL such as transfer outside of local data distributions.
arXiv Detail & Related papers (2020-12-15T19:30:29Z) - Federated Mutual Learning [65.46254760557073]
Federated Mutual Leaning (FML) allows clients training a generalized model collaboratively and a personalized model independently.
The experiments show that FML can achieve better performance than alternatives in typical Federated learning setting.
arXiv Detail & Related papers (2020-06-27T09:35:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.