FedZKT: Zero-Shot Knowledge Transfer towards Heterogeneous On-Device
Models in Federated Learning
- URL: http://arxiv.org/abs/2109.03775v1
- Date: Wed, 8 Sep 2021 16:53:07 GMT
- Title: FedZKT: Zero-Shot Knowledge Transfer towards Heterogeneous On-Device
Models in Federated Learning
- Authors: Lan Zhang, Xiaoyong Yuan
- Abstract summary: Federated learning enables distributed devices to learn a shared prediction model without centralizing on-device training data.
This paper proposes a new framework supporting federated learning across heterogeneous on-device models via Zero-shot Knowledge Transfer.
- Score: 6.9573683028565885
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning enables distributed devices to collaboratively learn a
shared prediction model without centralizing on-device training data. Most of
the current algorithms require comparable individual efforts to train on-device
models with the same structure and size, impeding participation from
resource-constrained devices. Given the widespread yet heterogeneous devices
nowadays, this paper proposes a new framework supporting federated learning
across heterogeneous on-device models via Zero-shot Knowledge Transfer, named
by FedZKT. Specifically, FedZKT allows participating devices to independently
determine their on-device models. To transfer knowledge across on-device
models, FedZKT develops a zero-shot distillation approach contrary to certain
prior research based on a public dataset or a pre-trained data generator. To
utmostly reduce on-device workload, the resource-intensive distillation task is
assigned to the server, which constructs a generator to adversarially train
with the ensemble of the received heterogeneous on-device models. The distilled
central knowledge will then be sent back in the form of the corresponding
on-device model parameters, which can be easily absorbed at the device side.
Experimental studies demonstrate the effectiveness and the robustness of FedZKT
towards heterogeneous on-device models and challenging federated learning
scenarios, such as non-iid data distribution and straggler effects.
Related papers
- One-shot Federated Learning via Synthetic Distiller-Distillate Communication [63.89557765137003]
One-shot Federated learning (FL) is a powerful technology facilitating collaborative training of machine learning models in a single round of communication.
We propose FedSD2C, a novel and practical one-shot FL framework designed to address these challenges.
arXiv Detail & Related papers (2024-12-06T17:05:34Z) - subMFL: Compatiple subModel Generation for Federated Learning in Device Heterogenous Environment [0.2796197251957244]
Federated Learning (FL) is commonly used in systems with distributed and heterogeneous devices.
We propose a model compression approach that enables heterogeneous devices with varying computing capacities to participate in the FL process.
arXiv Detail & Related papers (2024-05-30T12:49:34Z) - DFRD: Data-Free Robustness Distillation for Heterogeneous Federated
Learning [20.135235291912185]
Federated Learning (FL) is a privacy-constrained decentralized machine learning paradigm.
We propose a new FL method (namely DFRD) to learn a robust global model in the data-heterogeneous and model-heterogeneous FL scenarios.
arXiv Detail & Related papers (2023-09-24T04:29:22Z) - Unsupervised anomalies detection in IIoT edge devices networks using
federated learning [0.0]
Federated learning(FL) as a distributed machine learning approach performs training of a machine learning model on the device that gathered the data itself.
In this paper, we leverage the benefits of FL and implemented Fedavg algorithm on a recent dataset that represent the modern IoT/ IIoT device networks.
We also evaluated some shortcomings of Fedavg such as unfairness that happens during the training when struggling devices do not participate for every stage of training.
arXiv Detail & Related papers (2023-08-23T14:53:38Z) - Tackling Computational Heterogeneity in FL: A Few Theoretical Insights [68.8204255655161]
We introduce and analyse a novel aggregation framework that allows for formalizing and tackling computational heterogeneous data.
Proposed aggregation algorithms are extensively analyzed from a theoretical, and an experimental prospective.
arXiv Detail & Related papers (2023-07-12T16:28:21Z) - Federated Split GANs [12.007429155505767]
We propose an alternative approach to train ML models in user's devices themselves.
We focus on GANs (generative adversarial networks) and leverage their inherent privacy-preserving attribute.
Our system preserves data privacy, keeps a short training time, and yields same accuracy of model training in unconstrained devices.
arXiv Detail & Related papers (2022-07-04T23:53:47Z) - FedHiSyn: A Hierarchical Synchronous Federated Learning Framework for
Resource and Data Heterogeneity [56.82825745165945]
Federated Learning (FL) enables training a global model without sharing the decentralized raw data stored on multiple devices to protect data privacy.
We propose a hierarchical synchronous FL framework, i.e., FedHiSyn, to tackle the problems of straggler effects and outdated models.
We evaluate the proposed framework based on MNIST, EMNIST, CIFAR10 and CIFAR100 datasets and diverse heterogeneous settings of devices.
arXiv Detail & Related papers (2022-06-21T17:23:06Z) - Federated Learning with Downlink Device Selection [92.14944020945846]
We study federated edge learning, where a global model is trained collaboratively using privacy-sensitive data at the edge of a wireless network.
A parameter server (PS) keeps track of the global model and shares it with the wireless edge devices for training using their private local data.
We consider device selection based on downlink channels over which the PS shares the global model with the devices.
arXiv Detail & Related papers (2021-07-07T22:42:39Z) - Rethinking Architecture Design for Tackling Data Heterogeneity in
Federated Learning [53.73083199055093]
We show that attention-based architectures (e.g., Transformers) are fairly robust to distribution shifts.
Our experiments show that replacing convolutional networks with Transformers can greatly reduce catastrophic forgetting of previous devices.
arXiv Detail & Related papers (2021-06-10T21:04:18Z) - Federated Learning-based Active Authentication on Mobile Devices [98.23904302910022]
User active authentication on mobile devices aims to learn a model that can correctly recognize the enrolled user based on device sensor information.
We propose a novel user active authentication training, termed as Federated Active Authentication (FAA)
We show that existing FL/SL methods are suboptimal for FAA as they rely on the data to be distributed homogeneously.
arXiv Detail & Related papers (2021-04-14T22:59:08Z) - Fast-Convergent Federated Learning [82.32029953209542]
Federated learning is a promising solution for distributing machine learning tasks through modern networks of mobile devices.
We propose a fast-convergent federated learning algorithm, called FOLB, which performs intelligent sampling of devices in each round of model training.
arXiv Detail & Related papers (2020-07-26T14:37:51Z) - Evaluation Framework For Large-scale Federated Learning [10.127616622630514]
Federated learning is proposed as a machine learning setting to enable distributed edge devices, such as mobile phones, to collaboratively learn a shared prediction model.
In this paper, we introduce a framework designed for large-scale federated learning which consists of approaches to generating dataset and modular evaluation framework.
arXiv Detail & Related papers (2020-03-03T15:12:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.