Domain Discrepancy Aware Distillation for Model Aggregation in Federated
Learning
- URL: http://arxiv.org/abs/2210.02190v1
- Date: Tue, 4 Oct 2022 04:08:16 GMT
- Title: Domain Discrepancy Aware Distillation for Model Aggregation in Federated
Learning
- Authors: Shangchao Su and Bin Li and Xiangyang Xue
- Abstract summary: We describe two challenges, server-to-client discrepancy and client-to-client discrepancy, brought to the aggregation model by the domain discrepancies.
We propose an adaptive knowledge aggregation algorithm FedD3A based on domain discrepancy aware distillation to lower the bound.
- Score: 47.87639746826555
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Knowledge distillation has recently become popular as a method of model
aggregation on the server for federated learning. It is generally assumed that
there are abundant public unlabeled data on the server. However, in reality,
there exists a domain discrepancy between the datasets of the server domain and
a client domain, which limits the performance of knowledge distillation. How to
improve the aggregation under such a domain discrepancy setting is still an
open problem. In this paper, we first analyze the generalization bound of the
aggregation model produced from knowledge distillation for the client domains,
and then describe two challenges, server-to-client discrepancy and
client-to-client discrepancy, brought to the aggregation model by the domain
discrepancies. Following our analysis, we propose an adaptive knowledge
aggregation algorithm FedD3A based on domain discrepancy aware distillation to
lower the bound. FedD3A performs adaptive weighting at the sample level in each
round of FL. For each sample in the server domain, only the client models of
its similar domains will be selected for playing the teacher role. To achieve
this, we show that the discrepancy between the server-side sample and the
client domain can be approximately measured using a subspace projection matrix
calculated on each client without accessing its raw data. The server can thus
leverage the projection matrices from multiple clients to assign weights to the
corresponding teacher models for each server-side sample. We validate FedD3A on
two popular cross-domain datasets and show that it outperforms the compared
competitors in both cross-silo and cross-device FL settings.
Related papers
- Heterogeneous Federated Learning Using Knowledge Codistillation [23.895665011884102]
We propose a method that involves training a small model on the entire pool and a larger model on a subset of clients with higher capacity.
The models exchange information bidirectionally via knowledge distillation, utilizing an unlabeled dataset on a server without sharing parameters.
arXiv Detail & Related papers (2023-10-04T03:17:26Z) - Towards Instance-adaptive Inference for Federated Learning [80.38701896056828]
Federated learning (FL) is a distributed learning paradigm that enables multiple clients to learn a powerful global model by aggregating local training.
In this paper, we present a novel FL algorithm, i.e., FedIns, to handle intra-client data heterogeneity by enabling instance-adaptive inference in the FL framework.
Our experiments show that our FedIns outperforms state-of-the-art FL algorithms, e.g., a 6.64% improvement against the top-performing method with less than 15% communication cost on Tiny-ImageNet.
arXiv Detail & Related papers (2023-08-11T09:58:47Z) - HyperFed: Hyperbolic Prototypes Exploration with Consistent Aggregation
for Non-IID Data in Federated Learning [14.503047600805436]
Federated learning (FL) collaboratively models user data in a decentralized way.
In the real world, non-identical and independent data distributions (non-IID) among clients hinder the performance of FL due to three issues, i.e., the class statistics shifting, (2) the insufficient hierarchical information utilization, and (3) the inconsistency in aggregating clients.
arXiv Detail & Related papers (2023-07-26T02:43:38Z) - Efficient Distribution Similarity Identification in Clustered Federated
Learning via Principal Angles Between Client Data Subspaces [59.33965805898736]
Clustered learning has been shown to produce promising results by grouping clients into clusters.
Existing FL algorithms are essentially trying to group clients together with similar distributions.
Prior FL algorithms attempt similarities indirectly during training.
arXiv Detail & Related papers (2022-09-21T17:37:54Z) - Cross-domain Federated Object Detection [43.66352018668227]
Federated learning can enable multi-party collaborative learning without leaking client data.
We propose a cross-domain federated object detection framework, named FedOD.
arXiv Detail & Related papers (2022-06-30T03:09:59Z) - Game of Gradients: Mitigating Irrelevant Clients in Federated Learning [3.2095659532757916]
Federated learning (FL) deals with multiple clients participating in collaborative training of a machine learning model under the orchestration of a central server.
In this setup, each client's data is private to itself and is not transferable to other clients or the server.
We refer to these problems as Federated Relevant Client Selection (FRCS)
arXiv Detail & Related papers (2021-10-23T16:34:42Z) - Federated Multi-Target Domain Adaptation [99.93375364579484]
Federated learning methods enable us to train machine learning models on distributed user data while preserving its privacy.
We consider a more practical scenario where the distributed client data is unlabeled, and a centralized labeled dataset is available on the server.
We propose an effective DualAdapt method to address the new challenges.
arXiv Detail & Related papers (2021-08-17T17:53:05Z) - A Bayesian Federated Learning Framework with Online Laplace
Approximation [144.7345013348257]
Federated learning allows multiple clients to collaboratively learn a globally shared model.
We propose a novel FL framework that uses online Laplace approximation to approximate posteriors on both the client and server side.
We achieve state-of-the-art results on several benchmarks, clearly demonstrating the advantages of the proposed method.
arXiv Detail & Related papers (2021-02-03T08:36:58Z) - Federated Unsupervised Representation Learning [56.715917111878106]
We formulate a new problem in federated learning called Federated Unsupervised Representation Learning (FURL) to learn a common representation model without supervision.
FedCA is composed of two key modules: dictionary module to aggregate the representations of samples from each client and share with all clients for consistency of representation space and alignment module to align the representation of each client on a base model trained on a public data.
arXiv Detail & Related papers (2020-10-18T13:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.