Decentralized Federated Policy Gradient with Byzantine Fault-Tolerance
and Provably Fast Convergence
- URL: http://arxiv.org/abs/2401.03489v1
- Date: Sun, 7 Jan 2024 14:06:06 GMT
- Title: Decentralized Federated Policy Gradient with Byzantine Fault-Tolerance
and Provably Fast Convergence
- Authors: Philip Jordan, Florian Gr\"otschla, Flint Xiaofeng Fan, Roger
Wattenhofer
- Abstract summary: In Federated Reinforcement Learning (FRL), agents aim to collaboratively learn a common task, while each agent is acting in its local environment without exchanging raw trajectories.
We first propose a new centralized Byzantine fault-tolerant policy (PG) algorithm that improves existing methods by relying only on assumptions standard for non-fault-tolerant PG.
- Score: 21.935405256685307
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In Federated Reinforcement Learning (FRL), agents aim to collaboratively
learn a common task, while each agent is acting in its local environment
without exchanging raw trajectories. Existing approaches for FRL either (a) do
not provide any fault-tolerance guarantees (against misbehaving agents), or (b)
rely on a trusted central agent (a single point of failure) for aggregating
updates. We provide the first decentralized Byzantine fault-tolerant FRL
method. Towards this end, we first propose a new centralized Byzantine
fault-tolerant policy gradient (PG) algorithm that improves over existing
methods by relying only on assumptions standard for non-fault-tolerant PG.
Then, as our main contribution, we show how a combination of robust aggregation
and Byzantine-resilient agreement methods can be leveraged in order to
eliminate the need for a trusted central entity. Since our results represent
the first sample complexity analysis for Byzantine fault-tolerant decentralized
federated non-convex optimization, our technical contributions may be of
independent interest. Finally, we corroborate our theoretical results
experimentally for common RL environments, demonstrating the speed-up of
decentralized federations w.r.t. the number of participating agents and
resilience against various Byzantine attacks.
Related papers
- Boosting the Performance of Decentralized Federated Learning via Catalyst Acceleration [66.43954501171292]
We introduce Catalyst Acceleration and propose an acceleration Decentralized Federated Learning algorithm called DFedCata.
DFedCata consists of two main components: the Moreau envelope function, which addresses parameter inconsistencies, and Nesterov's extrapolation step, which accelerates the aggregation phase.
Empirically, we demonstrate the advantages of the proposed algorithm in both convergence speed and generalization performance on CIFAR10/100 with various non-iid data distributions.
arXiv Detail & Related papers (2024-10-09T06:17:16Z) - Byzantine-Robust Aggregation for Securing Decentralized Federated
Learning [0.32985979395737774]
Federated Learning (FL) emerges as a distributed machine learning approach that addresses privacy concerns by training AI models locally on devices.
Decentralized Federated Learning (DFL) extends the FL paradigm by eliminating the central server, thereby enhancing scalability and robustness through the avoidance of a single point of failure.
We present a novel Byzantine-robust aggregation algorithm to enhance the security of DFL environments, coined WFAgg.
arXiv Detail & Related papers (2024-09-26T11:36:08Z) - VALID: a Validated Algorithm for Learning in Decentralized Networks with Possible Adversarial Presence [13.612214163974459]
We introduce the paradigm of validated decentralized learning for undirected networks with heterogeneous data.
VALID protocol is the first to achieve a validated learning guarantee.
Remarkably, VALID retains optimal performance metrics in adversary-free environments.
arXiv Detail & Related papers (2024-05-12T15:55:43Z) - Byzantine-Resilient Federated PCA and Low Rank Column-wise Sensing [17.243528378512778]
This work considers two related learning problems in a federated attack prone setting: federated principal components analysis (PCA) and federated low rank column-wise sensing (LRCS)
The node attacks are assumed to be Byzantine which means that the attackers are omniscient and can collude.
We introduce a novel provably Byzantine-resilient communication-efficient and sampleefficient algorithm, called Subspace-Median, that solves the PCA problem and is a key part of the solution for the LRCS problem.
arXiv Detail & Related papers (2023-09-25T20:21:11Z) - Byzantine-Robust Online and Offline Distributed Reinforcement Learning [60.970950468309056]
We consider a distributed reinforcement learning setting where multiple agents explore the environment and communicate their experiences through a central server.
$alpha$-fraction of agents are adversarial and can report arbitrary fake information.
We seek to identify a near-optimal policy for the underlying Markov decision process in the presence of these adversarial agents.
arXiv Detail & Related papers (2022-06-01T00:44:53Z) - Monotonic Improvement Guarantees under Non-stationarity for
Decentralized PPO [66.5384483339413]
We present a new monotonic improvement guarantee for optimizing decentralized policies in cooperative Multi-Agent Reinforcement Learning (MARL)
We show that a trust region constraint can be effectively enforced in a principled way by bounding independent ratios based on the number of agents in training.
arXiv Detail & Related papers (2022-01-31T20:39:48Z) - Fault-Tolerant Federated Reinforcement Learning with Theoretical
Guarantee [25.555844784263236]
We propose the first Federated Reinforcement Learning framework that is tolerant to less than half of the participating agents being random system failures or adversarial attackers.
All theoretical results are empirically verified on various RL benchmark tasks.
arXiv Detail & Related papers (2021-10-26T23:01:22Z) - Dealing with Non-Stationarity in Multi-Agent Reinforcement Learning via
Trust Region Decomposition [52.06086375833474]
Non-stationarity is one thorny issue in multi-agent reinforcement learning.
We introduce a $delta$-stationarity measurement to explicitly model the stationarity of a policy sequence.
We propose a trust region decomposition network based on message passing to estimate the joint policy divergence.
arXiv Detail & Related papers (2021-02-21T14:46:50Z) - F2A2: Flexible Fully-decentralized Approximate Actor-critic for
Cooperative Multi-agent Reinforcement Learning [110.35516334788687]
Decentralized multi-agent reinforcement learning algorithms are sometimes unpractical in complicated applications.
We propose a flexible fully decentralized actor-critic MARL framework, which can handle large-scale general cooperative multi-agent setting.
Our framework can achieve scalability and stability for large-scale environment and reduce information transmission.
arXiv Detail & Related papers (2020-04-17T14:56:29Z) - Byzantine-resilient Decentralized Stochastic Gradient Descent [85.15773446094576]
We present an in-depth study towards the Byzantine resilience of decentralized learning systems.
We propose UBAR, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance.
arXiv Detail & Related papers (2020-02-20T05:11:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.