Quantization Robust Federated Learning for Efficient Inference on
Heterogeneous Devices
- URL: http://arxiv.org/abs/2206.10844v1
- Date: Wed, 22 Jun 2022 05:11:44 GMT
- Title: Quantization Robust Federated Learning for Efficient Inference on
Heterogeneous Devices
- Authors: Kartik Gupta, Marios Fournarakis, Matthias Reisser, Christos Louizos,
Markus Nagel
- Abstract summary: Federated Learning (FL) is a paradigm to distributively learn machine learning models from decentralized data that remains on-device.
We introduce multiple variants of federated averaging algorithm that train neural networks robust to quantization.
Our results demonstrate that integrating quantization robustness results in FL models that are significantly more robust to different bit-widths during quantized on-device inference.
- Score: 18.1568276196989
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated Learning (FL) is a machine learning paradigm to distributively
learn machine learning models from decentralized data that remains on-device.
Despite the success of standard Federated optimization methods, such as
Federated Averaging (FedAvg) in FL, the energy demands and hardware induced
constraints for on-device learning have not been considered sufficiently in the
literature. Specifically, an essential demand for on-device learning is to
enable trained models to be quantized to various bit-widths based on the energy
needs and heterogeneous hardware designs across the federation. In this work,
we introduce multiple variants of federated averaging algorithm that train
neural networks robust to quantization. Such networks can be quantized to
various bit-widths with only limited reduction in full precision model
accuracy. We perform extensive experiments on standard FL benchmarks to
evaluate our proposed FedAvg variants for quantization robustness and provide a
convergence analysis for our Quantization-Aware variants in FL. Our results
demonstrate that integrating quantization robustness results in FL models that
are significantly more robust to different bit-widths during quantized
on-device inference.
Related papers
- Prioritizing Modalities: Flexible Importance Scheduling in Federated Multimodal Learning [5.421492821020181]
Federated Learning (FL) is a distributed machine learning approach that enables devices to collaboratively train models without sharing their local data.
Applying FL to real-world data presents challenges, particularly as most existing FL research focuses on unimodal data.
We propose FlexMod, a novel approach to enhance computational efficiency in MFL by adaptively allocating training resources for each modality encoder.
arXiv Detail & Related papers (2024-08-13T01:14:27Z) - OnDev-LCT: On-Device Lightweight Convolutional Transformers towards
federated learning [29.798780069556074]
Federated learning (FL) has emerged as a promising approach to collaboratively train machine learning models across multiple edge devices.
We propose OnDev-LCT: Lightweight Convolutional Transformers for On-Device vision tasks with limited training data and resources.
arXiv Detail & Related papers (2024-01-22T02:17:36Z) - Mixed-Precision Quantization for Federated Learning on
Resource-Constrained Heterogeneous Devices [17.56259695496955]
We present a novel FL algorithm, FedMPQ, which introduces mixed-precision quantization to resource-heterogeneous FL systems.
Specifically, local models, quantized so as to satisfy bit-width constraint, are trained by optimizing an objective function.
To initialize the next round of local training, the server relies on the information learned in the previous training round to customize bit-width assignments of the models delivered to different clients.
arXiv Detail & Related papers (2023-11-29T22:43:40Z) - Semi-Federated Learning: Convergence Analysis and Optimization of A
Hybrid Learning Framework [70.83511997272457]
We propose a semi-federated learning (SemiFL) paradigm to leverage both the base station (BS) and devices for a hybrid implementation of centralized learning (CL) and FL.
We propose a two-stage algorithm to solve this intractable problem, in which we provide the closed-form solutions to the beamformers.
arXiv Detail & Related papers (2023-10-04T03:32:39Z) - Performance Optimization for Variable Bitwidth Federated Learning in
Wireless Networks [103.22651843174471]
This paper considers improving wireless communication and computation efficiency in federated learning (FL) via model quantization.
In the proposed bitwidth FL scheme, edge devices train and transmit quantized versions of their local FL model parameters to a coordinating server, which aggregates them into a quantized global model and synchronizes the devices.
We show that the FL training process can be described as a Markov decision process and propose a model-based reinforcement learning (RL) method to optimize action selection over iterations.
arXiv Detail & Related papers (2022-09-21T08:52:51Z) - Efficient Split-Mix Federated Learning for On-Demand and In-Situ
Customization [107.72786199113183]
Federated learning (FL) provides a distributed learning framework for multiple participants to collaborate learning without sharing raw data.
In this paper, we propose a novel Split-Mix FL strategy for heterogeneous participants that, once training is done, provides in-situ customization of model sizes and robustness.
arXiv Detail & Related papers (2022-03-18T04:58:34Z) - Bitwidth Heterogeneous Federated Learning with Progressive Weight
Dequantization [58.31288475660333]
We introduce a pragmatic Federated Learning scenario with bitwidth Heterogeneous Federated Learning (BHFL)
BHFL brings in a new challenge, that the aggregation of model parameters with different bitwidths could result in severe performance degeneration.
We propose ProWD framework, which has a trainable weight dequantizer at the central server that progressively reconstructs the low-bitwidth weights into higher bitwidth weights, and finally into full-precision weights.
arXiv Detail & Related papers (2022-02-23T12:07:02Z) - Fast-Convergent Federated Learning [82.32029953209542]
Federated learning is a promising solution for distributing machine learning tasks through modern networks of mobile devices.
We propose a fast-convergent federated learning algorithm, called FOLB, which performs intelligent sampling of devices in each round of model training.
arXiv Detail & Related papers (2020-07-26T14:37:51Z) - UVeQFed: Universal Vector Quantization for Federated Learning [179.06583469293386]
Federated learning (FL) is an emerging approach to train such learning models without requiring the users to share their possibly private labeled data.
In FL, each user trains its copy of the learning model locally. The server then collects the individual updates and aggregates them into a global model.
We show that combining universal vector quantization methods with FL yields a decentralized training system in which the compression of the trained models induces only a minimum distortion.
arXiv Detail & Related papers (2020-06-05T07:10:22Z) - Ternary Compression for Communication-Efficient Federated Learning [17.97683428517896]
Federated learning provides a potential solution to privacy-preserving and secure machine learning.
We propose a ternary federated averaging protocol (T-FedAvg) to reduce the upstream and downstream communication of federated learning systems.
Our results show that the proposed T-FedAvg is effective in reducing communication costs and can even achieve slightly better performance on non-IID data.
arXiv Detail & Related papers (2020-03-07T11:55:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.