Hierarchical Quantized Federated Learning: Convergence Analysis and
System Design
- URL: http://arxiv.org/abs/2103.14272v1
- Date: Fri, 26 Mar 2021 05:48:36 GMT
- Title: Hierarchical Quantized Federated Learning: Convergence Analysis and
System Design
- Authors: Lumin Liu, Jun Zhang, Shenghui Song, Khaled B. Letaief
- Abstract summary: Federated learning is a collaborative machine to train deep neural networks without clients' private data.
Previous works assume one central parameter either at the cloud or at the edge.
This paper exploits the advantages of both cloud servers and considers Hierarchical Quantized Federated Learning system.
- Score: 7.481427303081613
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning is a collaborative machine learning framework to train
deep neural networks without accessing clients' private data. Previous works
assume one central parameter server either at the cloud or at the edge. A cloud
server can aggregate knowledge from all participating clients but suffers high
communication overhead and latency, while an edge server enjoys more efficient
communications during model update but can only reach a limited number of
clients. This paper exploits the advantages of both cloud and edge servers and
considers a Hierarchical Quantized Federated Learning (HQFL) system with one
cloud server, several edge servers and many clients, adopting a
communication-efficient training algorithm, Hier-Local-QSGD. The high
communication efficiency comes from frequent local aggregations at the edge
servers and fewer aggregations at the cloud server, as well as weight
quantization during model uploading. A tight convergence bound for non-convex
objective loss functions is derived, which is then applied to investigate two
design problems, namely, the accuracy-latency trade-off and edge-client
association. It will be shown that given a latency budget for the whole
training process, there is an optimal parameter choice with respect to the two
aggregation intervals and two quantization levels. For the edge-client
association problem, it is found that the edge-client association strategy has
no impact on the convergence speed. Empirical simulations shall verify the
findings from the convergence analysis and demonstrate the accuracy-latency
trade-off in the hierarchical federated learning system.
Related papers
- Communication Efficient ConFederated Learning: An Event-Triggered SAGA
Approach [67.27031215756121]
Federated learning (FL) is a machine learning paradigm that targets model training without gathering the local data over various data sources.
Standard FL, which employs a single server, can only support a limited number of users, leading to degraded learning capability.
In this work, we consider a multi-server FL framework, referred to as emphConfederated Learning (CFL) in order to accommodate a larger number of users.
arXiv Detail & Related papers (2024-02-28T03:27:10Z) - Timely Asynchronous Hierarchical Federated Learning: Age of Convergence [59.96266198512243]
We consider an asynchronous hierarchical federated learning setting with a client-edge-cloud framework.
The clients exchange the trained parameters with their corresponding edge servers, which update the locally aggregated model.
The goal of each client is to converge to the global model, while maintaining timeliness of the clients.
arXiv Detail & Related papers (2023-06-21T17:39:16Z) - Adaptive Federated Pruning in Hierarchical Wireless Networks [69.6417645730093]
Federated Learning (FL) is a privacy-preserving distributed learning framework where a server aggregates models updated by multiple devices without accessing their private datasets.
In this paper, we introduce model pruning for HFL in wireless networks to reduce the neural network scale.
We show that our proposed HFL with model pruning achieves similar learning accuracy compared with the HFL without model pruning and reduces about 50 percent communication cost.
arXiv Detail & Related papers (2023-05-15T22:04:49Z) - An Efficient Split Fine-tuning Framework for Edge and Cloud
Collaborative Learning [20.118073642453034]
We design an efficient split fine-tuning framework for edge and cloud collaborative learning.
We compress the intermediate output of a neural network to reduce the communication volume between the edge device and the cloud server.
Our framework can reduce the communication traffic by 96 times with little impact on the model accuracy.
arXiv Detail & Related papers (2022-11-30T02:55:21Z) - Scalable Collaborative Learning via Representation Sharing [53.047460465980144]
Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
arXiv Detail & Related papers (2022-11-20T10:49:22Z) - Optimizing Server-side Aggregation For Robust Federated Learning via
Subspace Training [80.03567604524268]
Non-IID data distribution across clients and poisoning attacks are two main challenges in real-world federated learning systems.
We propose SmartFL, a generic approach that optimize the server-side aggregation process.
We provide theoretical analyses of the convergence and generalization capacity for SmartFL.
arXiv Detail & Related papers (2022-11-10T13:20:56Z) - Time Minimization in Hierarchical Federated Learning [11.678121177730718]
Federated learning is a modern decentralized machine learning technique where user equipments perform machine learning tasks locally and then upload the model parameters to a central server.
In this paper, we consider a 3-layer hierarchical federated learning system which involves model parameter exchanges between the cloud and edge servers.
arXiv Detail & Related papers (2022-10-07T13:53:20Z) - Joint Parameter-and-Bandwidth Allocation for Improving the Efficiency of
Partitioned Edge Learning [73.82875010696849]
Machine learning algorithms are deployed at the network edge for training artificial intelligence (AI) models.
This paper focuses on the novel joint design of parameter (computation load) allocation and bandwidth allocation.
arXiv Detail & Related papers (2020-03-10T05:52:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.