$\texttt{FedBC}$: Calibrating Global and Local Models via Federated
Learning Beyond Consensus
- URL: http://arxiv.org/abs/2206.10815v1
- Date: Wed, 22 Jun 2022 02:42:04 GMT
- Title: $\texttt{FedBC}$: Calibrating Global and Local Models via Federated
Learning Beyond Consensus
- Authors: Amrit Singh Bedi, Chen Fan, Alec Koppel, Anit Kumar Sahu, Brian M.
Sadler, Furong Huang, and Dinesh Manocha
- Abstract summary: In federated learning (FL), the objective of collaboratively learning a global model through aggregation of model updates across devices tends to oppose the goal of personalization via local information.
In this work, we calibrate this tradeoff in a quantitative manner through a multi-criterion-based optimization.
We demonstrate that $texttFedBC$ balances the global and local model test accuracy metrics across a suite datasets.
- Score: 66.62731854746856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In federated learning (FL), the objective of collaboratively learning a
global model through aggregation of model updates across devices tends to
oppose the goal of personalization via local information. In this work, we
calibrate this tradeoff in a quantitative manner through a multi-criterion
optimization-based framework, which we cast as a constrained program: the
objective for a device is its local objective, which it seeks to minimize while
satisfying nonlinear constraints that quantify the proximity between the local
and the global model. By considering the Lagrangian relaxation of this problem,
we develop an algorithm that allows each node to minimize its local component
of Lagrangian through queries to a first-order gradient oracle. Then, the
server executes Lagrange multiplier ascent steps followed by a Lagrange
multiplier-weighted averaging step. We call this instantiation of the
primal-dual method Federated Learning Beyond Consensus ($\texttt{FedBC}$).
Theoretically, we establish that $\texttt{FedBC}$ converges to a first-order
stationary point at rates that matches the state of the art, up to an
additional error term that depends on the tolerance parameter that arises due
to the proximity constraints. Overall, the analysis is a novel characterization
of primal-dual methods applied to non-convex saddle point problems with
nonlinear constraints. Finally, we demonstrate that $\texttt{FedBC}$ balances
the global and local model test accuracy metrics across a suite of datasets
(Synthetic, MNIST, CIFAR-10, Shakespeare), achieving competitive performance
with the state of the art.
Related papers
- FedHPL: Efficient Heterogeneous Federated Learning with Prompt Tuning and Logit Distillation [32.305134875959226]
Federated learning (FL) is a privacy-preserving paradigm that enables distributed clients to collaboratively train models with a central server.
We propose FedHPL, a parameter-efficient unified $textbfFed$erated learning framework for $textbfH$eterogeneous settings.
We show that our framework outperforms state-of-the-art FL approaches, with less overhead and training rounds.
arXiv Detail & Related papers (2024-05-27T15:25:32Z) - Calibrated One Round Federated Learning with Bayesian Inference in the
Predictive Space [27.259110269667826]
Federated Learning (FL) involves training a model over a dataset distributed among clients.
Small and noisy datasets are common, highlighting the need for well-calibrated models.
We propose $beta$-Predictive Bayes, a Bayesian FL algorithm that interpolates between a mixture and product of the predictive posteriors.
arXiv Detail & Related papers (2023-12-15T14:17:16Z) - DFedADMM: Dual Constraints Controlled Model Inconsistency for
Decentralized Federated Learning [52.83811558753284]
Decentralized learning (DFL) discards the central server and establishes a decentralized communication network.
Existing DFL methods still suffer from two major challenges: local inconsistency and local overfitting.
arXiv Detail & Related papers (2023-08-16T11:22:36Z) - Constrained Optimization via Exact Augmented Lagrangian and Randomized
Iterative Sketching [55.28394191394675]
We develop an adaptive inexact Newton method for equality-constrained nonlinear, nonIBS optimization problems.
We demonstrate the superior performance of our method on benchmark nonlinear problems, constrained logistic regression with data from LVM, and a PDE-constrained problem.
arXiv Detail & Related papers (2023-05-28T06:33:37Z) - FedSpeed: Larger Local Interval, Less Communication Round, and Higher
Generalization Accuracy [84.45004766136663]
Federated learning is an emerging distributed machine learning framework.
It suffers from the non-vanishing biases introduced by the local inconsistent optimal and the rugged client-drifts by the local over-fitting.
We propose a novel and practical method, FedSpeed, to alleviate the negative impacts posed by these problems.
arXiv Detail & Related papers (2023-02-21T03:55:29Z) - Online Federated Learning via Non-Stationary Detection and Adaptation
amidst Concept Drift [39.12903814606534]
Federated Learning (FL) is an emerging domain in the broader context of artificial intelligence research.
Existing literature in FL mostly assumes stationary data generation processes.
We introduce a multiscale algorithmic framework which combines theoretical guarantees of textitFedAvg and textitFedOMD algorithms in near stationary settings.
arXiv Detail & Related papers (2022-11-22T20:45:48Z) - Robust One Round Federated Learning with Predictive Space Bayesian
Inference [19.533268415744338]
We show how the global predictive posterior can be approximated using client predictive posteriors.
We present an algorithm based on this idea, which performs MCMC sampling at each client to obtain an estimate of the local posterior, and then aggregates these in one round to obtain a global ensemble model.
arXiv Detail & Related papers (2022-06-20T01:06:59Z) - A Bayesian Federated Learning Framework with Online Laplace
Approximation [144.7345013348257]
Federated learning allows multiple clients to collaboratively learn a globally shared model.
We propose a novel FL framework that uses online Laplace approximation to approximate posteriors on both the client and server side.
We achieve state-of-the-art results on several benchmarks, clearly demonstrating the advantages of the proposed method.
arXiv Detail & Related papers (2021-02-03T08:36:58Z) - Faster Non-Convex Federated Learning via Global and Local Momentum [57.52663209739171]
textttFedGLOMO is the first (first-order) FLtexttFedGLOMO algorithm.
Our algorithm is provably optimal even with communication between the clients and the server.
arXiv Detail & Related papers (2020-12-07T21:05:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.