Adaptive Quantization of Model Updates for Communication-Efficient
Federated Learning
- URL: http://arxiv.org/abs/2102.04487v1
- Date: Mon, 8 Feb 2021 19:14:21 GMT
- Title: Adaptive Quantization of Model Updates for Communication-Efficient
Federated Learning
- Authors: Divyansh Jhunjhunwala, Advait Gadhikar, Gauri Joshi, Yonina C. Eldar
- Abstract summary: Communication of model updates between client nodes and the central aggregating server is a major bottleneck in federated learning.
Gradient quantization is an effective way of reducing the number of bits required to communicate each model update.
We propose an adaptive quantization strategy called AdaFL that aims to achieve communication efficiency as well as a low error floor.
- Score: 75.45968495410047
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Communication of model updates between client nodes and the central
aggregating server is a major bottleneck in federated learning, especially in
bandwidth-limited settings and high-dimensional models. Gradient quantization
is an effective way of reducing the number of bits required to communicate each
model update, albeit at the cost of having a higher error floor due to the
higher variance of the stochastic gradients. In this work, we propose an
adaptive quantization strategy called AdaQuantFL that aims to achieve
communication efficiency as well as a low error floor by changing the number of
quantization levels during the course of training. Experiments on training deep
neural networks show that our method can converge in much fewer communicated
bits as compared to fixed quantization level setups, with little or no impact
on training and test accuracy.
Related papers
- FedAQ: Communication-Efficient Federated Edge Learning via Joint Uplink and Downlink Adaptive Quantization [11.673528138087244]
Federated learning (FL) is a powerful machine learning paradigm which leverages the data as well as the computational resources of clients, while protecting clients' data privacy.
Previous research has primarily focused on the uplink communication, employing either fixed-bit quantization or adaptive quantization methods.
In this work, we introduce a holistic approach by joint uplink and downlink adaptive quantization to reduce the communication overhead.
arXiv Detail & Related papers (2024-06-26T08:14:23Z) - Quantized Adaptive Subgradient Algorithms and Their Applications [39.103587572626026]
We propose quantized composite mirror descent adaptive subgradient (QCMD adagrad) and quantized regularized dual average adaptive subgradient (QRDA adagrad) for distributed training.
A quantized gradient-based adaptive learning rate matrix is constructed to achieve a balance between communication costs, accuracy, and model sparsity.
arXiv Detail & Related papers (2022-08-11T04:04:03Z) - Fundamental Limits of Communication Efficiency for Model Aggregation in
Distributed Learning: A Rate-Distortion Approach [54.311495894129585]
We study the limit of communication cost of model aggregation in distributed learning from a rate-distortion perspective.
It is found that the communication gain by exploiting the correlation between worker nodes is significant for SignSGD.
arXiv Detail & Related papers (2022-06-28T13:10:40Z) - FedDQ: Communication-Efficient Federated Learning with Descending
Quantization [5.881154276623056]
Federated learning (FL) is an emerging privacy-preserving distributed learning scheme.
FL suffers from critical communication bottleneck due to large model size and frequent model aggregation.
This paper proposes an opposite approach to do adaptive quantization.
arXiv Detail & Related papers (2021-10-05T18:56:28Z) - Cluster-Promoting Quantization with Bit-Drop for Minimizing Network
Quantization Loss [61.26793005355441]
Cluster-Promoting Quantization (CPQ) finds the optimal quantization grids for neural networks.
DropBits is a new bit-drop technique that revises the standard dropout regularization to randomly drop bits instead of neurons.
We experimentally validate our method on various benchmark datasets and network architectures.
arXiv Detail & Related papers (2021-09-05T15:15:07Z) - All at Once Network Quantization via Collaborative Knowledge Transfer [56.95849086170461]
We develop a novel collaborative knowledge transfer approach for efficiently training the all-at-once quantization network.
Specifically, we propose an adaptive selection strategy to choose a high-precision enquoteteacher for transferring knowledge to the low-precision student.
To effectively transfer knowledge, we develop a dynamic block swapping method by randomly replacing the blocks in the lower-precision student network with the corresponding blocks in the higher-precision teacher network.
arXiv Detail & Related papers (2021-03-02T03:09:03Z) - CosSGD: Nonlinear Quantization for Communication-efficient Federated
Learning [62.65937719264881]
Federated learning facilitates learning across clients without transferring local data on these clients to a central server.
We propose a nonlinear quantization for compressed gradient descent, which can be easily utilized in federated learning.
Our system significantly reduces the communication cost by up to three orders of magnitude, while maintaining convergence and accuracy of the training process.
arXiv Detail & Related papers (2020-12-15T12:20:28Z) - Optimal Gradient Quantization Condition for Communication-Efficient
Distributed Training [99.42912552638168]
Communication of gradients is costly for training deep neural networks with multiple devices in computer vision applications.
In this work, we deduce the optimal condition of both the binary and multi-level gradient quantization for textbfANY gradient distribution.
Based on the optimal condition, we develop two novel quantization schemes: biased BinGrad and unbiased ORQ for binary and multi-level gradient quantization respectively.
arXiv Detail & Related papers (2020-02-25T18:28:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.