Characterization of the Global Bias Problem in Aerial Federated Learning
- URL: http://arxiv.org/abs/2212.14360v1
- Date: Thu, 29 Dec 2022 16:19:36 GMT
- Title: Characterization of the Global Bias Problem in Aerial Federated Learning
- Authors: Ruslan Zhagypar, Nour Kouzayha, Hesham ElSawy, Hayssam Dahrouj, and
Tareq Y. Al-Naffouri
- Abstract summary: Unmanned aerial vehicles (UAVs) mobility enables flexible and customized federated learning (FL) at the network edge.
The distribution of the global model and the aggregation of the local updates within the FL learning rounds at the UAVs are governed by the reliability of the wireless channel.
This paper characterizes the global bias problem of aerial FL in large-scale UAV networks.
- Score: 30.555678487200456
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unmanned aerial vehicles (UAVs) mobility enables flexible and customized
federated learning (FL) at the network edge. However, the underlying
uncertainties in the aerial-terrestrial wireless channel may lead to a biased
FL model. In particular, the distribution of the global model and the
aggregation of the local updates within the FL learning rounds at the UAVs are
governed by the reliability of the wireless channel. This creates an
undesirable bias towards the training data of ground devices with better
channel conditions, and vice versa. This paper characterizes the global bias
problem of aerial FL in large-scale UAV networks. To this end, the paper
proposes a channel-aware distribution and aggregation scheme to enforce equal
contribution from all devices in the FL training as a means to resolve the
global bias problem. We demonstrate the convergence of the proposed method by
experimenting with the MNIST dataset and show its superiority compared to
existing methods. The obtained results enable system parameter tuning to
relieve the impact of the aerial channel deficiency on the FL convergence rate.
Related papers
- UAV-assisted Unbiased Hierarchical Federated Learning: Performance and Convergence Analysis [16.963596661873954]
Hierarchical federated learning (HFL) is a key paradigm to distribute learning across edge devices to reach global intelligence.
In HFL, each edge device trains a local model using its respective data and transmits the updated model parameters to an edge server for local aggregation.
This paper proposes an unbiased HFL algorithm for unmanned aerial vehicle (UAV)-assisted wireless networks.
arXiv Detail & Related papers (2024-07-05T06:23:01Z) - Decentralized Federated Learning Over Imperfect Communication Channels [68.08499874460857]
This paper analyzes the impact of imperfect communication channels on decentralized federated learning (D-FL)
It determines the optimal number of local aggregations per training round, adapting to the network topology and imperfect channels.
It is seen that D-FL, with an optimal number of local aggregations, can outperform its potential alternatives by over 10% in training accuracy.
arXiv Detail & Related papers (2024-05-21T16:04:32Z) - Over-the-Air Federated Learning and Optimization [52.5188988624998]
We focus on Federated learning (FL) via edge-the-air computation (AirComp)
We describe the convergence of AirComp-based FedAvg (AirFedAvg) algorithms under both convex and non- convex settings.
For different types of local updates that can be transmitted by edge devices (i.e., model, gradient, model difference), we reveal that transmitting in AirFedAvg may cause an aggregation error.
In addition, we consider more practical signal processing schemes to improve the communication efficiency and extend the convergence analysis to different forms of model aggregation error caused by these signal processing schemes.
arXiv Detail & Related papers (2023-10-16T05:49:28Z) - A Federated Channel Modeling System using Generative Neural Networks [8.424767096311928]
The paper proposes a data-driven approach to air-to-ground channel estimation in a millimeter-wave wireless network on an unmanned aerial vehicle.
We use Federated Learning (FL) for channel estimation and can predict the air-to-ground path loss between a low-altitude platform and a terrestrial terminal.
To evaluate the effectiveness of the proposed model, we evaluate its performance using Kullback-Leibler divergence (KL), and Wasserstein distance between the synthetic data distribution generated by the model and the actual data distribution.
arXiv Detail & Related papers (2023-05-30T08:50:22Z) - Random Orthogonalization for Federated Learning in Massive MIMO Systems [85.71432283670114]
We propose a novel communication design for federated learning (FL) in a massive multiple-input and multiple-output (MIMO) wireless system.
Key novelty of random orthogonalization comes from the tight coupling of FL and two unique characteristics of massive MIMO -- channel hardening and favorable propagation.
We extend this principle to the downlink communication phase and develop a simple but highly effective model broadcast method for FL.
arXiv Detail & Related papers (2022-10-18T14:17:10Z) - Over-the-Air Federated Learning from Heterogeneous Data [107.05618009955094]
Federated learning (FL) is a framework for distributed learning of centralized models.
We develop a Convergent OTA FL (COTAF) algorithm which enhances the common local gradient descent (SGD) FL algorithm.
We numerically show that the precoding induced by COTAF notably improves the convergence rate and the accuracy of models trained via OTA FL.
arXiv Detail & Related papers (2020-09-27T08:28:25Z) - Federated Learning in the Sky: Joint Power Allocation and Scheduling
with UAV Swarms [98.78553146823829]
Unmanned aerial vehicle (UAV) swarms must exploit machine learning (ML) in order to execute various tasks.
In this paper, a novel framework is proposed to implement distributed learning (FL) algorithms within a UAV swarm.
arXiv Detail & Related papers (2020-02-19T14:04:01Z) - Convergence Time Optimization for Federated Learning over Wireless
Networks [160.82696473996566]
A wireless network is considered in which wireless users transmit their local FL models (trained using their locally collected data) to a base station (BS)
The BS, acting as a central controller, generates a global FL model using the received local FL models and broadcasts it back to all users.
Due to the limited number of resource blocks (RBs) in a wireless network, only a subset of users can be selected to transmit their local FL model parameters to the BS.
Since each user has unique training data samples, the BS prefers to include all local user FL models to generate a converged global FL model.
arXiv Detail & Related papers (2020-01-22T01:55:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.