Over-the-Air Federated Edge Learning with Hierarchical Clustering
- URL: http://arxiv.org/abs/2207.09232v1
- Date: Tue, 19 Jul 2022 12:42:12 GMT
- Title: Over-the-Air Federated Edge Learning with Hierarchical Clustering
- Authors: Ozan Ayg\"un, Mohammad Kazemi, Deniz G\"und\"uz, Tolga M. Duman
- Abstract summary: In over-the-air (OTA) aggregation, mobile users (MUs) aim to reach a consensus on a global model with the help of a parameter server (PS)
In OTA FL, MUs train their models using local data at every training round and transmit their gradients simultaneously using the same frequency band in an uncoded fashion.
While the OTA FL has a significantly decreased communication cost, it is susceptible to adverse channel effects and noise.
We propose a wireless-based hierarchical FL scheme that uses intermediate servers (ISs) to form clusters at the areas where the MUs are more densely located
- Score: 21.51594138166343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We examine federated learning (FL) with over-the-air (OTA) aggregation, where
mobile users (MUs) aim to reach a consensus on a global model with the help of
a parameter server (PS) that aggregates the local gradients. In OTA FL, MUs
train their models using local data at every training round and transmit their
gradients simultaneously using the same frequency band in an uncoded fashion.
Based on the received signal of the superposed gradients, the PS performs a
global model update. While the OTA FL has a significantly decreased
communication cost, it is susceptible to adverse channel effects and noise.
Employing multiple antennas at the receiver side can reduce these effects, yet
the path-loss is still a limiting factor for users located far away from the
PS. To ameliorate this issue, in this paper, we propose a wireless-based
hierarchical FL scheme that uses intermediate servers (ISs) to form clusters at
the areas where the MUs are more densely located. Our scheme utilizes OTA
cluster aggregations for the communication of the MUs with their corresponding
IS, and OTA global aggregations from the ISs to the PS. We present a
convergence analysis for the proposed algorithm, and show through numerical
evaluations of the derived analytical expressions and experimental results that
utilizing ISs results in a faster convergence and a better performance than the
OTA FL alone while using less transmit power. We also validate the results on
the performance using different number of cluster iterations with different
datasets and data distributions. We conclude that the best choice of cluster
aggregations depends on the data distribution among the MUs and the clusters.
Related papers
- Enhancing Federated Learning Convergence with Dynamic Data Queue and Data Entropy-driven Participant Selection [13.825031686864559]
Federated Learning (FL) is a decentralized approach for collaborative model training on edge devices.
We present a method to improve convergence in FL by creating a global subset of data on the server and dynamically distributing it across devices.
Our approach results in a substantial accuracy boost of approximately 5% for the MNIST dataset, around 18% for CIFAR-10, and 20% for CIFAR-100 with a 10% global subset of data, outperforming the state-of-the-art (SOTA) aggregation algorithms.
arXiv Detail & Related papers (2024-10-23T11:47:04Z) - Communication Efficient ConFederated Learning: An Event-Triggered SAGA
Approach [67.27031215756121]
Federated learning (FL) is a machine learning paradigm that targets model training without gathering the local data over various data sources.
Standard FL, which employs a single server, can only support a limited number of users, leading to degraded learning capability.
In this work, we consider a multi-server FL framework, referred to as emphConfederated Learning (CFL) in order to accommodate a larger number of users.
arXiv Detail & Related papers (2024-02-28T03:27:10Z) - Exploiting Label Skews in Federated Learning with Model Concatenation [39.38427550571378]
Federated Learning (FL) has emerged as a promising solution to perform deep learning on different data owners without exchanging raw data.
Among different non-IID types, label skews have been challenging and common in image classification and other tasks.
We propose FedConcat, a simple and effective approach that degrades these local models as the base of the global model.
arXiv Detail & Related papers (2023-12-11T10:44:52Z) - Over-the-Air Federated Learning and Optimization [52.5188988624998]
We focus on Federated learning (FL) via edge-the-air computation (AirComp)
We describe the convergence of AirComp-based FedAvg (AirFedAvg) algorithms under both convex and non- convex settings.
For different types of local updates that can be transmitted by edge devices (i.e., model, gradient, model difference), we reveal that transmitting in AirFedAvg may cause an aggregation error.
In addition, we consider more practical signal processing schemes to improve the communication efficiency and extend the convergence analysis to different forms of model aggregation error caused by these signal processing schemes.
arXiv Detail & Related papers (2023-10-16T05:49:28Z) - Adaptive Federated Pruning in Hierarchical Wireless Networks [69.6417645730093]
Federated Learning (FL) is a privacy-preserving distributed learning framework where a server aggregates models updated by multiple devices without accessing their private datasets.
In this paper, we introduce model pruning for HFL in wireless networks to reduce the neural network scale.
We show that our proposed HFL with model pruning achieves similar learning accuracy compared with the HFL without model pruning and reduces about 50 percent communication cost.
arXiv Detail & Related papers (2023-05-15T22:04:49Z) - FLCC: Efficient Distributed Federated Learning on IoMT over CSMA/CA [0.0]
Federated Learning (FL) has emerged as a promising approach for privacy preservation.
This article investigates the performance of FL on an application that might be used to improve a remote healthcare system over ad hoc networks.
We present two metrics to evaluate the network performance: 1) probability of successful transmission while minimizing the interference, and 2) performance of distributed FL model in terms of accuracy and loss.
arXiv Detail & Related papers (2023-03-29T16:36:42Z) - Hierarchical Over-the-Air Federated Edge Learning [26.553214197361626]
We propose hierarchical over-the-air federated learning (HOTAFL) to form clusters near mobile users (MUs)
We demonstrate through theoretical and experimental results that local aggregation in each cluster before global aggregation leads to a better performance and faster convergence than OTA FL.
arXiv Detail & Related papers (2021-12-21T13:02:10Z) - Bayesian Federated Learning over Wireless Networks [87.37301441859925]
Federated learning is a privacy-preserving and distributed training method using heterogeneous data sets stored at local devices.
This paper presents an efficient modified BFL algorithm called scalableBFL (SBFL)
arXiv Detail & Related papers (2020-12-31T07:32:44Z) - Over-the-Air Federated Learning from Heterogeneous Data [107.05618009955094]
Federated learning (FL) is a framework for distributed learning of centralized models.
We develop a Convergent OTA FL (COTAF) algorithm which enhances the common local gradient descent (SGD) FL algorithm.
We numerically show that the precoding induced by COTAF notably improves the convergence rate and the accuracy of models trained via OTA FL.
arXiv Detail & Related papers (2020-09-27T08:28:25Z) - Federated Learning in the Sky: Joint Power Allocation and Scheduling
with UAV Swarms [98.78553146823829]
Unmanned aerial vehicle (UAV) swarms must exploit machine learning (ML) in order to execute various tasks.
In this paper, a novel framework is proposed to implement distributed learning (FL) algorithms within a UAV swarm.
arXiv Detail & Related papers (2020-02-19T14:04:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.