Federated Learning in the Presence of Adversarial Client Unavailability
- URL: http://arxiv.org/abs/2305.19971v2
- Date: Tue, 20 Feb 2024 03:38:31 GMT
- Title: Federated Learning in the Presence of Adversarial Client Unavailability
- Authors: Lili Su, Ming Xiang, Jiaming Xu, Pengkun Yang
- Abstract summary: Federated learning is a decentralized machine learning framework that enables collaborative model without revealing raw data.
Due to the diverse hardware software limitations, a client may not always be available for the computation requests from the server.
In harsh environments like battlefields, adversaries can selectively silence specific clients.
- Score: 16.201377650598516
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Federated learning is a decentralized machine learning framework that enables
collaborative model training without revealing raw data. Due to the diverse
hardware and software limitations, a client may not always be available for the
computation requests from the parameter server. An emerging line of research is
devoted to tackling arbitrary client unavailability. However, existing work
still imposes structural assumptions on the unavailability patterns, impeding
their applicability in challenging scenarios wherein the unavailability
patterns are beyond the control of the parameter server. Moreover, in harsh
environments like battlefields, adversaries can selectively and adaptively
silence specific clients. In this paper, we relax the structural assumptions
and consider adversarial client unavailability. To quantify the degrees of
client unavailability, we use the notion of $\epsilon$-adversary dropout
fraction. We show that simple variants of FedAvg or FedProx, albeit completely
agnostic to $\epsilon$, converge to an estimation error on the order of
$\epsilon (G^2 + \sigma^2)$ for non-convex global objectives and $\epsilon(G^2
+ \sigma^2)/\mu^2$ for $\mu$ strongly convex global objectives, where $G$ is a
heterogeneity parameter and $\sigma^2$ is the noise level. Conversely, we prove
that any algorithm has to suffer an estimation error of at least $\epsilon (G^2
+ \sigma^2)/8$ and $\epsilon(G^2 + \sigma^2)/(8\mu^2)$ for non-convex global
objectives and $\mu$-strongly convex global objectives. Furthermore, the
convergence speeds of the FedAvg or FedProx variants are $O(1/\sqrt{T})$ for
non-convex objectives and $O(1/T)$ for strongly-convex objectives, both of
which are the best possible for any first-order method that only has access to
noisy gradients.
Related papers
- Robust Model Evaluation over Large-scale Federated Networks [8.700087812420687]
We address the challenge of certifying the performance of a machine learning model on an unseen target network.
We derive theoretical guarantees for the model's empirical average loss and provide uniform bounds on the risk CDF.
Our bounds are computable in time with a number of queries to the $K$ clients, preserving client privacy by querying only the model's loss on private data.
arXiv Detail & Related papers (2024-10-26T18:45:15Z) - Federated Linear Bandits with Finite Adversarial Actions [20.1041278044797]
We study a federated linear bandits model, where $M$ clients communicate with a central server to solve a linear contextual bandits problem.
To address the unique challenges of adversarial finite action sets, we propose the FedSupLinUCB algorithm.
We prove that FedSupLinUCB achieves a total regret of $tildeO(sqrtd T)$, where $T$ is the total number of arm pulls from all clients, and $d$ is the ambient dimension of the linear model.
arXiv Detail & Related papers (2023-11-02T03:41:58Z) - Addressing GAN Training Instabilities via Tunable Classification Losses [8.151943266391493]
Generative adversarial networks (GANs) allow generating synthetic data with formal guarantees.
We show that all symmetric $f$-divergences are equivalent in convergence.
We also highlight the value of tuning $(alpha_D,alpha_G)$ in alleviating training instabilities for the synthetic 2D Gaussian mixture ring.
arXiv Detail & Related papers (2023-10-27T17:29:07Z) - A Unified Framework for Uniform Signal Recovery in Nonlinear Generative
Compressed Sensing [68.80803866919123]
Under nonlinear measurements, most prior results are non-uniform, i.e., they hold with high probability for a fixed $mathbfx*$ rather than for all $mathbfx*$ simultaneously.
Our framework accommodates GCS with 1-bit/uniformly quantized observations and single index models as canonical examples.
We also develop a concentration inequality that produces tighter bounds for product processes whose index sets have low metric entropy.
arXiv Detail & Related papers (2023-09-25T17:54:19Z) - Towards Bias Correction of FedAvg over Nonuniform and Time-Varying
Communications [26.597515045714502]
Federated learning (FL) is a decentralized learning framework wherein a parameter server (PS) and a collection of clients collaboratively train a model via a global objective.
We show that when the channel conditions are heterogeneous across clients are changing over time, the FedFederated Postponed global model fails to postpone the gossip-type information mixing errors.
arXiv Detail & Related papers (2023-06-01T01:52:03Z) - The Sample Complexity of Online Contract Design [120.9833763323407]
We study the hidden-action principal-agent problem in an online setting.
In each round, the principal posts a contract that specifies the payment to the agent based on each outcome.
The agent then makes a strategic choice of action that maximizes her own utility, but the action is not directly observable by the principal.
arXiv Detail & Related papers (2022-11-10T17:59:42Z) - Faster Rates for Compressed Federated Learning with Client-Variance
Reduction [23.169998435268504]
We show that COFIG and FRECON converge within $O(frac(1+omega)sqrtNSepsilon2)$ communication rounds.
In the convex setting COFIG converges within $O(frac(1+omega)sqrtNSepsilon2)$ communication rounds, which, to the best of our knowledge, is the first result.
arXiv Detail & Related papers (2021-12-24T16:28:18Z) - Safe Learning under Uncertain Objectives and Constraints [66.05180398174286]
We consider non-textitunknown yet safety-critical optimization problems under textitunknown yet safety-critical constraints.
Such problems naturally arise in a variety of domains including robotics, manufacturing, and medical procedures.
A crucial component of our analysis is to introduce and apply a technique called shrinkage in the context of safe optimization.
arXiv Detail & Related papers (2020-06-23T20:51:00Z) - Agnostic Learning of a Single Neuron with Gradient Descent [92.7662890047311]
We consider the problem of learning the best-fitting single neuron as measured by the expected square loss.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
arXiv Detail & Related papers (2020-05-29T07:20:35Z) - A Simple Convergence Proof of Adam and Adagrad [74.24716715922759]
We show a proof of convergence between the Adam Adagrad and $O(d(N)/st)$ algorithms.
Adam converges with the same convergence $O(d(N)/st)$ when used with the default parameters.
arXiv Detail & Related papers (2020-03-05T01:56:17Z) - Curse of Dimensionality on Randomized Smoothing for Certifiable
Robustness [151.67113334248464]
We show that extending the smoothing technique to defend against other attack models can be challenging.
We present experimental results on CIFAR to validate our theory.
arXiv Detail & Related papers (2020-02-08T22:02:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.