Online Learning with Adversaries: A Differential-Inclusion Analysis
- URL: http://arxiv.org/abs/2304.01525v2
- Date: Tue, 26 Sep 2023 08:58:27 GMT
- Title: Online Learning with Adversaries: A Differential-Inclusion Analysis
- Authors: Swetha Ganesh, Alexandre Reiffers-Masson, Gugan Thoppe
- Abstract summary: We introduce an observation-matrix-based framework for fully asynchronous online Federated Learning with adversaries.
Our main result is that the proposed algorithm almost surely converges to the desired mean $mu.$
We derive this convergence using a novel differential-inclusion-based two-timescale analysis.
- Score: 52.43460995467893
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce an observation-matrix-based framework for fully asynchronous
online Federated Learning (FL) with adversaries. In this work, we demonstrate
its effectiveness in estimating the mean of a random vector. Our main result is
that the proposed algorithm almost surely converges to the desired mean $\mu.$
This makes ours the first asynchronous FL method to have an a.s. convergence
guarantee in the presence of adversaries. We derive this convergence using a
novel differential-inclusion-based two-timescale analysis. Two other highlights
of our proof include (a) the use of a novel Lyapunov function to show that
$\mu$ is the unique global attractor for our algorithm's limiting dynamics, and
(b) the use of martingale and stopping-time theory to show that our algorithm's
iterates are almost surely bounded.
Related papers
- Adaptive, Doubly Optimal No-Regret Learning in Strongly Monotone and Exp-Concave Games with Gradient Feedback [75.29048190099523]
Online gradient descent (OGD) is well known to be doubly optimal under strong convexity or monotonicity assumptions.
In this paper, we design a fully adaptive OGD algorithm, textsfAdaOGD, that does not require a priori knowledge of these parameters.
arXiv Detail & Related papers (2023-10-21T18:38:13Z) - Solving Constrained Variational Inequalities via an Interior Point
Method [88.39091990656107]
We develop an interior-point approach to solve constrained variational inequality (cVI) problems.
We provide convergence guarantees for ACVI in two general classes of problems.
Unlike previous work in this setting, ACVI provides a means to solve cVIs when the constraints are nontrivial.
arXiv Detail & Related papers (2022-06-21T17:55:13Z) - A Lyapunov Theory for Finite-Sample Guarantees of Asynchronous
Q-Learning and TD-Learning Variants [39.28675942566465]
This paper develops a framework to study finite-sample convergence guarantees of a class of value-based asynchronous RL algorithms.
As a by-product, we provide theoretical insights into the bias-variance trade-off, i.e., efficiency of bootstrapping in RL.
arXiv Detail & Related papers (2021-02-02T15:48:19Z) - Learning Sign-Constrained Support Vector Machines [0.24466725954625884]
We develop two optimization algorithms for minimizing the empirical risk under the sign constraints.
One of the two algorithms is based on the projected gradient method, in which each iteration of the projected gradient method takes $O(nd)$ computational cost.
We empirically demonstrate that the sign constraints are a promising technique when similarities to the training examples compose the feature vector.
arXiv Detail & Related papers (2021-01-05T12:08:17Z) - Byzantine-Resilient Non-Convex Stochastic Gradient Descent [61.6382287971982]
adversary-resilient distributed optimization, in which.
machines can independently compute gradients, and cooperate.
Our algorithm is based on a new concentration technique, and its sample complexity.
It is very practical: it improves upon the performance of all prior methods when no.
setting machines are present.
arXiv Detail & Related papers (2020-12-28T17:19:32Z) - Finite-Time Analysis for Double Q-learning [50.50058000948908]
We provide the first non-asymptotic, finite-time analysis for double Q-learning.
We show that both synchronous and asynchronous double Q-learning are guaranteed to converge to an $epsilon$-accurate neighborhood of the global optimum.
arXiv Detail & Related papers (2020-09-29T18:48:21Z) - A Distributional Analysis of Sampling-Based Reinforcement Learning
Algorithms [67.67377846416106]
We present a distributional approach to theoretical analyses of reinforcement learning algorithms for constant step-sizes.
We show that value-based methods such as TD($lambda$) and $Q$-Learning have update rules which are contractive in the space of distributions of functions.
arXiv Detail & Related papers (2020-03-27T05:13:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.