Aggregating Capacity in FL through Successive Layer Training for
Computationally-Constrained Devices
- URL: http://arxiv.org/abs/2305.17005v2
- Date: Mon, 27 Nov 2023 11:43:53 GMT
- Title: Aggregating Capacity in FL through Successive Layer Training for
Computationally-Constrained Devices
- Authors: Kilian Pfeiffer, Ramin Khalili, J\"org Henkel
- Abstract summary: Federated learning (FL) is usually performed on resource-constrained edge devices.
FL training process should be adjusted to such constraints.
We propose a new method that enables successive freezing and training of the parameters of the FL model at devices.
- Score: 3.4530027457862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning (FL) is usually performed on resource-constrained edge
devices, e.g., with limited memory for the computation. If the required memory
to train a model exceeds this limit, the device will be excluded from the
training. This can lead to a lower accuracy as valuable data and computation
resources are excluded from training, also causing bias and unfairness. The FL
training process should be adjusted to such constraints. The state-of-the-art
techniques propose training subsets of the FL model at constrained devices,
reducing their resource requirements for training. But these techniques largely
limit the co-adaptation among parameters of the model and are highly
inefficient, as we show: it is actually better to train a smaller (less
accurate) model by the system where all the devices can train the model
end-to-end, than applying such techniques. We propose a new method that enables
successive freezing and training of the parameters of the FL model at devices,
reducing the training's resource requirements at the devices, while still
allowing enough co-adaptation between parameters. We show through extensive
experimental evaluation that our technique greatly improves the accuracy of the
trained model (by 52.4 p.p.) compared with the state of the art, efficiently
aggregating the computation capacity available on distributed devices.
Related papers
- Efficient Federated Finetuning of Tiny Transformers with Resource-Constrained Devices [10.676390348161888]
Large Language Models (LLMs) require massive amounts of data for training and induce high resource requirements.
To fine-tune such a model in a parameter-efficient way, techniques like Adapter or LoRA have been developed.
We show that our presented scheme outperforms the current state of the art when dealing with homogeneous or heterogeneous computation and memory constraints.
arXiv Detail & Related papers (2024-11-12T14:22:16Z) - Efficient Asynchronous Federated Learning with Sparsification and
Quantization [55.6801207905772]
Federated Learning (FL) is attracting more and more attention to collaboratively train a machine learning model without transferring raw data.
FL generally exploits a parameter server and a large number of edge devices during the whole process of the model training.
We propose TEASQ-Fed to exploit edge devices to asynchronously participate in the training process by actively applying for tasks.
arXiv Detail & Related papers (2023-12-23T07:47:07Z) - Adaptive Model Pruning and Personalization for Federated Learning over
Wireless Networks [72.59891661768177]
Federated learning (FL) enables distributed learning across edge devices while protecting data privacy.
We consider a FL framework with partial model pruning and personalization to overcome these challenges.
This framework splits the learning model into a global part with model pruning shared with all devices to learn data representations and a personalized part to be fine-tuned for a specific device.
arXiv Detail & Related papers (2023-09-04T21:10:45Z) - Vertical Federated Learning over Cloud-RAN: Convergence Analysis and
System Optimization [82.12796238714589]
We propose a novel cloud radio access network (Cloud-RAN) based vertical FL system to enable fast and accurate model aggregation.
We characterize the convergence behavior of the vertical FL algorithm considering both uplink and downlink transmissions.
We establish a system optimization framework by joint transceiver and fronthaul quantization design, for which successive convex approximation and alternate convex search based system optimization algorithms are developed.
arXiv Detail & Related papers (2023-05-04T09:26:03Z) - Performance Optimization for Variable Bitwidth Federated Learning in
Wireless Networks [103.22651843174471]
This paper considers improving wireless communication and computation efficiency in federated learning (FL) via model quantization.
In the proposed bitwidth FL scheme, edge devices train and transmit quantized versions of their local FL model parameters to a coordinating server, which aggregates them into a quantized global model and synchronizes the devices.
We show that the FL training process can be described as a Markov decision process and propose a model-based reinforcement learning (RL) method to optimize action selection over iterations.
arXiv Detail & Related papers (2022-09-21T08:52:51Z) - CoCoFL: Communication- and Computation-Aware Federated Learning via
Partial NN Freezing and Quantization [3.219812767529503]
We present a novel FL technique, CoCoFL, which maintains the full NN structure on all devices.
CoCoFL efficiently utilizes the available resources on devices and allows constrained devices to make a significant contribution to the FL system.
arXiv Detail & Related papers (2022-03-10T16:45:05Z) - Resource-Efficient and Delay-Aware Federated Learning Design under Edge
Heterogeneity [10.702853653891902]
Federated learning (FL) has emerged as a popular methodology for distributing machine learning across wireless edge devices.
In this work, we consider optimizing the tradeoff between model performance and resource utilization in FL.
Our proposed StoFedDelAv incorporates a localglobal model combiner into the FL computation step.
arXiv Detail & Related papers (2021-12-27T22:30:15Z) - FedHe: Heterogeneous Models and Communication-Efficient Federated
Learning [0.0]
Federated learning (FL) is able to manage edge devices to cooperatively train a model while maintaining the training data local and private.
We propose a novel FL method, called FedHe, inspired by knowledge distillation, which can train heterogeneous models and support asynchronous training processes.
arXiv Detail & Related papers (2021-10-19T12:18:37Z) - LCS: Learning Compressible Subspaces for Adaptive Network Compression at
Inference Time [57.52251547365967]
We propose a method for training a "compressible subspace" of neural networks that contains a fine-grained spectrum of models.
We present results for achieving arbitrarily fine-grained accuracy-efficiency trade-offs at inference time for structured and unstructured sparsity.
Our algorithm extends to quantization at variable bit widths, achieving accuracy on par with individually trained networks.
arXiv Detail & Related papers (2021-10-08T17:03:34Z) - Fast-Convergent Federated Learning [82.32029953209542]
Federated learning is a promising solution for distributing machine learning tasks through modern networks of mobile devices.
We propose a fast-convergent federated learning algorithm, called FOLB, which performs intelligent sampling of devices in each round of model training.
arXiv Detail & Related papers (2020-07-26T14:37:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.