FedNew: A Communication-Efficient and Privacy-Preserving Newton-Type
Method for Federated Learning
- URL: http://arxiv.org/abs/2206.08829v1
- Date: Fri, 17 Jun 2022 15:21:39 GMT
- Title: FedNew: A Communication-Efficient and Privacy-Preserving Newton-Type
Method for Federated Learning
- Authors: Anis Elgabli and Chaouki Ben Issaid and Amrit S. Bedi and Ketan
Rajawat and Mehdi Bennis and Vaneet Aggarwal
- Abstract summary: We introduce a novel framework called FedNew in which there is no need to transmit Hessian information from clients to PS.
FedNew hides the gradient information and results in a privacy-preserving approach compared to the existing state-of-the-art.
- Score: 75.46959684676371
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Newton-type methods are popular in federated learning due to their fast
convergence. Still, they suffer from two main issues, namely: low communication
efficiency and low privacy due to the requirement of sending Hessian
information from clients to parameter server (PS). In this work, we introduced
a novel framework called FedNew in which there is no need to transmit Hessian
information from clients to PS, hence resolving the bottleneck to improve
communication efficiency. In addition, FedNew hides the gradient information
and results in a privacy-preserving approach compared to the existing
state-of-the-art. The core novel idea in FedNew is to introduce a two level
framework, and alternate between updating the inverse Hessian-gradient product
using only one alternating direction method of multipliers (ADMM) step and then
performing the global model update using Newton's method. Though only one ADMM
pass is used to approximate the inverse Hessian-gradient product at each
iteration, we develop a novel theoretical approach to show the converging
behavior of FedNew for convex problems. Additionally, a significant reduction
in communication overhead is achieved by utilizing stochastic quantization.
Numerical results using real datasets show the superiority of FedNew compared
to existing methods in terms of communication costs.
Related papers
- FedNS: A Fast Sketching Newton-Type Algorithm for Federated Learning [27.957498393822338]
We introduce a novel approach to tackle this issue while still achieving fast convergence rates.
Our proposed method, named as Federated Newton Sketch methods (FedNS), approximates the centralized Newton's method by communicating the sketched square-root Hessian instead of the exact Hessian.
We provide convergence analysis based on statistical learning for the federated Newton sketch approaches.
arXiv Detail & Related papers (2024-01-05T10:06:41Z) - Personalized Federated Learning under Mixture of Distributions [98.25444470990107]
We propose a novel approach to Personalized Federated Learning (PFL), which utilizes Gaussian mixture models (GMM) to fit the input data distributions across diverse clients.
FedGMM possesses an additional advantage of adapting to new clients with minimal overhead, and it also enables uncertainty quantification.
Empirical evaluations on synthetic and benchmark datasets demonstrate the superior performance of our method in both PFL classification and novel sample detection.
arXiv Detail & Related papers (2023-05-01T20:04:46Z) - Magnitude Matters: Fixing SIGNSGD Through Magnitude-Aware Sparsification
in the Presence of Data Heterogeneity [60.791736094073]
Communication overhead has become one of the major bottlenecks in the distributed training of deep neural networks.
We propose a magnitude-driven sparsification scheme, which addresses the non-convergence issue of SIGNSGD.
The proposed scheme is validated through experiments on Fashion-MNIST, CIFAR-10, and CIFAR-100 datasets.
arXiv Detail & Related papers (2023-02-19T17:42:35Z) - Meta Knowledge Condensation for Federated Learning [65.20774786251683]
Existing federated learning paradigms usually extensively exchange distributed models at a central solver to achieve a more powerful model.
This would incur severe communication burden between a server and multiple clients especially when data distributions are heterogeneous.
Unlike existing paradigms, we introduce an alternative perspective to significantly decrease the communication cost in federate learning.
arXiv Detail & Related papers (2022-09-29T15:07:37Z) - Communication-Compressed Adaptive Gradient Method for Distributed
Nonconvex Optimization [21.81192774458227]
One of the major bottlenecks is the large communication cost between the central server and the local workers.
Our proposed distributed learning framework features an effective gradient gradient compression strategy.
arXiv Detail & Related papers (2021-11-01T04:54:55Z) - Federated Learning via Plurality Vote [38.778944321534084]
Federated learning allows collaborative workers to solve a machine learning problem while preserving data privacy.
Recent studies have tackled various challenges in federated learning.
We propose a new scheme named federated learning via plurality vote (FedVote)
arXiv Detail & Related papers (2021-10-06T18:16:22Z) - Enhancing the Generalization for Intent Classification and Out-of-Domain
Detection in SLU [70.44344060176952]
Intent classification is a major task in spoken language understanding (SLU)
Recent works have shown that using extra data and labels can improve the OOD detection performance.
This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection.
arXiv Detail & Related papers (2021-06-28T08:27:38Z) - Distributed Second Order Methods with Fast Rates and Compressed
Communication [6.069611493148631]
We develop several new communication-efficient second-order methods for distributed optimization.
We prove global sublinear and linear convergence rates, and a fast superlinear rate.
Our results are supported with experimental results on real datasets.
arXiv Detail & Related papers (2021-02-14T14:06:45Z) - TaylorGAN: Neighbor-Augmented Policy Update for Sample-Efficient Natural
Language Generation [79.4205462326301]
TaylorGAN is a novel approach to score function-based natural language generation.
It augments the gradient estimation by off-policy update and the first-order Taylor expansion.
It enables us to train NLG models from scratch with smaller batch size.
arXiv Detail & Related papers (2020-11-27T02:26:15Z) - Improving Federated Relational Data Modeling via Basis Alignment and
Weight Penalty [18.096788806121754]
Federated learning (FL) has attracted increasing attention in recent years.
We present a modified version of the graph neural network algorithm that performs federated modeling over Knowledge Graph (KG)
We propose a novel optimization algorithm, named FedAlign, with 1) optimal transportation (OT) for on-client personalization and 2) weight constraint to speed up the convergence.
Empirical results show that our proposed method outperforms the state-of-the-art FL methods, such as FedAVG and FedProx, with better convergence.
arXiv Detail & Related papers (2020-11-23T12:52:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.