Energy and Spectrum Efficient Federated Learning via High-Precision
Over-the-Air Computation
- URL: http://arxiv.org/abs/2208.07237v1
- Date: Mon, 15 Aug 2022 14:47:21 GMT
- Title: Energy and Spectrum Efficient Federated Learning via High-Precision
Over-the-Air Computation
- Authors: Liang Li, Chenpei Huang, Dian Shi, Hao Wang, Xiangwei Zhou, Minglei
Shu, Miao Pan
- Abstract summary: Federated learning (FL) enables mobile devices to collaboratively learn a shared prediction model while keeping data locally.
There are two major research challenges to practically deploy FL over mobile devices.
We propose a novel multi-bit over-the-air computation (M-AirComp) approach for spectrum-efficient aggregation of local model updates in FL.
- Score: 26.499025986273832
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning (FL) enables mobile devices to collaboratively learn a
shared prediction model while keeping data locally. However, there are two
major research challenges to practically deploy FL over mobile devices: (i)
frequent wireless updates of huge size gradients v.s. limited spectrum
resources, and (ii) energy-hungry FL communication and local computing during
training v.s. battery-constrained mobile devices. To address those challenges,
in this paper, we propose a novel multi-bit over-the-air computation
(M-AirComp) approach for spectrum-efficient aggregation of local model updates
in FL and further present an energy-efficient FL design for mobile devices.
Specifically, a high-precision digital modulation scheme is designed and
incorporated in the M-AirComp, allowing mobile devices to upload model updates
at the selected positions simultaneously in the multi-access channel. Moreover,
we theoretically analyze the convergence property of our FL algorithm. Guided
by FL convergence analysis, we formulate a joint transmission probability and
local computing control optimization, aiming to minimize the overall energy
consumption (i.e., iterative local computing + multi-round communications) of
mobile devices in FL. Extensive simulation results show that our proposed
scheme outperforms existing ones in terms of spectrum utilization, energy
efficiency, and learning accuracy.
Related papers
- WHALE-FL: Wireless and Heterogeneity Aware Latency Efficient Federated Learning over Mobile Devices via Adaptive Subnetwork Scheduling [17.029433544096257]
We develop a wireless and aware latency efficient FL (WHALE-FL) approach to accelerate FL training through adaptive subnetwork scheduling.
Our evaluation shows that, compared with peer designs, WHALE-FL effectively accelerates FL training without sacrificing learning accuracy.
arXiv Detail & Related papers (2024-05-01T22:01:40Z) - Semi-Federated Learning: Convergence Analysis and Optimization of A
Hybrid Learning Framework [70.83511997272457]
We propose a semi-federated learning (SemiFL) paradigm to leverage both the base station (BS) and devices for a hybrid implementation of centralized learning (CL) and FL.
We propose a two-stage algorithm to solve this intractable problem, in which we provide the closed-form solutions to the beamformers.
arXiv Detail & Related papers (2023-10-04T03:32:39Z) - Adaptive Model Pruning and Personalization for Federated Learning over
Wireless Networks [72.59891661768177]
Federated learning (FL) enables distributed learning across edge devices while protecting data privacy.
We consider a FL framework with partial model pruning and personalization to overcome these challenges.
This framework splits the learning model into a global part with model pruning shared with all devices to learn data representations and a personalized part to be fine-tuned for a specific device.
arXiv Detail & Related papers (2023-09-04T21:10:45Z) - Vertical Federated Learning over Cloud-RAN: Convergence Analysis and
System Optimization [82.12796238714589]
We propose a novel cloud radio access network (Cloud-RAN) based vertical FL system to enable fast and accurate model aggregation.
We characterize the convergence behavior of the vertical FL algorithm considering both uplink and downlink transmissions.
We establish a system optimization framework by joint transceiver and fronthaul quantization design, for which successive convex approximation and alternate convex search based system optimization algorithms are developed.
arXiv Detail & Related papers (2023-05-04T09:26:03Z) - Automated Federated Learning in Mobile Edge Networks -- Fast Adaptation
and Convergence [83.58839320635956]
Federated Learning (FL) can be used in mobile edge networks to train machine learning models in a distributed manner.
Recent FL has been interpreted within a Model-Agnostic Meta-Learning (MAML) framework, which brings FL significant advantages in fast adaptation and convergence over heterogeneous datasets.
This paper addresses how much benefit MAML brings to FL and how to maximize such benefit over mobile edge networks.
arXiv Detail & Related papers (2023-03-23T02:42:10Z) - Digital Over-the-Air Federated Learning in Multi-Antenna Systems [30.137208705209627]
We study the performance optimization of federated learning (FL) over a realistic wireless communication system with digital modulation and over-the-air computation (AirComp)
We propose a modified federated averaging (FedAvg) algorithm that combines digital modulation with AirComp to mitigate wireless fading while ensuring the communication efficiency.
An artificial neural network (ANN) is used to estimate the local FL models of all devices and adjust the beamforming matrices at the PS for future model transmission.
arXiv Detail & Related papers (2023-02-04T07:26:06Z) - Performance Optimization for Variable Bitwidth Federated Learning in
Wireless Networks [103.22651843174471]
This paper considers improving wireless communication and computation efficiency in federated learning (FL) via model quantization.
In the proposed bitwidth FL scheme, edge devices train and transmit quantized versions of their local FL model parameters to a coordinating server, which aggregates them into a quantized global model and synchronizes the devices.
We show that the FL training process can be described as a Markov decision process and propose a model-based reinforcement learning (RL) method to optimize action selection over iterations.
arXiv Detail & Related papers (2022-09-21T08:52:51Z) - CFLIT: Coexisting Federated Learning and Information Transfer [18.30671838758503]
We study the coexistence of over-the-air FL and traditional information transfer (IT) in a mobile edge network.
We propose a coexisting federated learning and information transfer (CFLIT) communication framework, where the FL and IT devices share the wireless spectrum in an OFDM system.
arXiv Detail & Related papers (2022-07-26T13:17:28Z) - To Talk or to Work: Flexible Communication Compression for Energy
Efficient Federated Learning over Heterogeneous Mobile Edge Devices [78.38046945665538]
federated learning (FL) over massive mobile edge devices opens new horizons for numerous intelligent mobile applications.
FL imposes huge communication and computation burdens on participating devices due to periodical global synchronization and continuous local training.
We develop a convergence-guaranteed FL algorithm enabling flexible communication compression.
arXiv Detail & Related papers (2020-12-22T02:54:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.