Semi-Federated Learning: Convergence Analysis and Optimization of A
Hybrid Learning Framework
- URL: http://arxiv.org/abs/2310.02559v1
- Date: Wed, 4 Oct 2023 03:32:39 GMT
- Title: Semi-Federated Learning: Convergence Analysis and Optimization of A
Hybrid Learning Framework
- Authors: Jingheng Zheng, Wanli Ni, Hui Tian, Deniz Gunduz, Tony Q. S. Quek, Zhu
Han
- Abstract summary: We propose a semi-federated learning (SemiFL) paradigm to leverage both the base station (BS) and devices for a hybrid implementation of centralized learning (CL) and FL.
We propose a two-stage algorithm to solve this intractable problem, in which we provide the closed-form solutions to the beamformers.
- Score: 70.83511997272457
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Under the organization of the base station (BS), wireless federated learning
(FL) enables collaborative model training among multiple devices. However, the
BS is merely responsible for aggregating local updates during the training
process, which incurs a waste of the computational resource at the BS. To
tackle this issue, we propose a semi-federated learning (SemiFL) paradigm to
leverage the computing capabilities of both the BS and devices for a hybrid
implementation of centralized learning (CL) and FL. Specifically, each device
sends both local gradients and data samples to the BS for training a shared
global model. To improve communication efficiency over the same time-frequency
resources, we integrate over-the-air computation for aggregation and
non-orthogonal multiple access for transmission by designing a novel
transceiver structure. To gain deep insights, we conduct convergence analysis
by deriving a closed-form optimality gap for SemiFL and extend the result to
two extra cases. In the first case, the BS uses all accumulated data samples to
calculate the CL gradient, while a decreasing learning rate is adopted in the
second case. Our analytical results capture the destructive effect of wireless
communication and show that both FL and CL are special cases of SemiFL. Then,
we formulate a non-convex problem to reduce the optimality gap by jointly
optimizing the transmit power and receive beamformers. Accordingly, we propose
a two-stage algorithm to solve this intractable problem, in which we provide
the closed-form solutions to the beamformers. Extensive simulation results on
two real-world datasets corroborate our theoretical analysis, and show that the
proposed SemiFL outperforms conventional FL and achieves 3.2% accuracy gain on
the MNIST dataset compared to state-of-the-art benchmarks.
Related papers
- Device Sampling and Resource Optimization for Federated Learning in Cooperative Edge Networks [17.637761046608]
Federated learning (FedL) distributes machine learning (ML) across worker devices by having them train local models that are periodically aggregated by a server.
FedL ignores two important characteristics of contemporary wireless networks: (i) the network may contain heterogeneous communication/computation resources, and (ii) there may be significant overlaps in devices' local data distributions.
We develop a novel optimization methodology that jointly accounts for these factors via intelligent device sampling complemented by device-to-device (D2D) offloading.
arXiv Detail & Related papers (2023-11-07T21:17:59Z) - Adaptive Model Pruning and Personalization for Federated Learning over
Wireless Networks [72.59891661768177]
Federated learning (FL) enables distributed learning across edge devices while protecting data privacy.
We consider a FL framework with partial model pruning and personalization to overcome these challenges.
This framework splits the learning model into a global part with model pruning shared with all devices to learn data representations and a personalized part to be fine-tuned for a specific device.
arXiv Detail & Related papers (2023-09-04T21:10:45Z) - Automated Federated Learning in Mobile Edge Networks -- Fast Adaptation
and Convergence [83.58839320635956]
Federated Learning (FL) can be used in mobile edge networks to train machine learning models in a distributed manner.
Recent FL has been interpreted within a Model-Agnostic Meta-Learning (MAML) framework, which brings FL significant advantages in fast adaptation and convergence over heterogeneous datasets.
This paper addresses how much benefit MAML brings to FL and how to maximize such benefit over mobile edge networks.
arXiv Detail & Related papers (2023-03-23T02:42:10Z) - Performance Optimization for Variable Bitwidth Federated Learning in
Wireless Networks [103.22651843174471]
This paper considers improving wireless communication and computation efficiency in federated learning (FL) via model quantization.
In the proposed bitwidth FL scheme, edge devices train and transmit quantized versions of their local FL model parameters to a coordinating server, which aggregates them into a quantized global model and synchronizes the devices.
We show that the FL training process can be described as a Markov decision process and propose a model-based reinforcement learning (RL) method to optimize action selection over iterations.
arXiv Detail & Related papers (2022-09-21T08:52:51Z) - Predictive GAN-powered Multi-Objective Optimization for Hybrid Federated
Split Learning [56.125720497163684]
We propose a hybrid federated split learning framework in wireless networks.
We design a parallel computing scheme for model splitting without label sharing, and theoretically analyze the influence of the delayed gradient caused by the scheme on the convergence speed.
arXiv Detail & Related papers (2022-09-02T10:29:56Z) - Device Sampling for Heterogeneous Federated Learning: Theory,
Algorithms, and Implementation [24.084053136210027]
We develop a sampling methodology based on graph sequential convolutional networks (GCNs)
We find that our methodology while sampling less than 5% of all devices outperforms conventional federated learning (FedL) substantially both in terms of trained model accuracy and required resource utilization.
arXiv Detail & Related papers (2021-01-04T05:59:50Z) - Federated Learning via Intelligent Reflecting Surface [30.935389187215474]
Over-the-air computation algorithm (AirComp) based learning (FL) is capable of achieving fast model aggregation by exploiting the waveform superposition property of multiple access channels.
In this paper, we propose a two-step optimization framework to achieve fast yet reliable model aggregation for AirComp-based FL.
Simulation results will demonstrate that our proposed framework and the deployment of an IRS can achieve a lower training loss and higher FL prediction accuracy than the baseline algorithms.
arXiv Detail & Related papers (2020-11-10T11:29:57Z) - Delay Minimization for Federated Learning Over Wireless Communication
Networks [172.42768672943365]
The problem of delay computation for federated learning (FL) over wireless communication networks is investigated.
A bisection search algorithm is proposed to obtain the optimal solution.
Simulation results show that the proposed algorithm can reduce delay by up to 27.3% compared to conventional FL methods.
arXiv Detail & Related papers (2020-07-05T19:00:07Z) - Federated Learning for Hybrid Beamforming in mm-Wave Massive MIMO [12.487990897680422]
We introduce a federated learning (FL) based framework for hybrid beamforming, where the model training is performed at the base station.
We design a convolutional neural network, in which the input is the channel data, yielding the analog beamformers at the output.
FL is demonstrated to be more tolerant to the imperfections and corruptions in the channel data as well as having less transmission overhead than CML.
arXiv Detail & Related papers (2020-05-20T11:21:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.