A Robust Classification Framework for Byzantine-Resilient Stochastic
Gradient Descent
- URL: http://arxiv.org/abs/2301.07498v1
- Date: Mon, 16 Jan 2023 10:40:09 GMT
- Title: A Robust Classification Framework for Byzantine-Resilient Stochastic
Gradient Descent
- Authors: Shashank Reddy Chirra, Kalyan Varma Nadimpalli, Shrisha Rao
- Abstract summary: This paper proposes a Robust Gradient Classification Framework (RGCF) for Byzantine fault tolerance in distributed gradient descent.
RGCF is not dependent on the number of workers; it can scale up to training instances with a large number of workers without a loss in performance.
- Score: 3.5450828190071655
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper proposes a Robust Gradient Classification Framework (RGCF) for
Byzantine fault tolerance in distributed stochastic gradient descent. The
framework consists of a pattern recognition filter which we train to be able to
classify individual gradients as Byzantine by using their direction alone. This
filter is robust to an arbitrary number of Byzantine workers for convex as well
as non-convex optimisation settings, which is a significant improvement on the
prior work that is robust to Byzantine faults only when up to 50% of the
workers are Byzantine. This solution does not require an estimate of the number
of Byzantine workers; its running time is not dependent on the number of
workers and can scale up to training instances with a large number of workers
without a loss in performance. We validate our solution by training
convolutional neural networks on the MNIST dataset in the presence of Byzantine
workers.
Related papers
- Dynamic Byzantine-Robust Learning: Adapting to Switching Byzantine Workers [10.632248569865236]
Byzantine-robust learning has emerged as a prominent fault-tolerant distributed machine learning framework.
We propose DynaBRO -- a new method capable of withstanding any sub-linear number of identity changes across rounds.
Our method utilizes a multi-level Monte Carlo gradient technique applied at the server to robustly aggregated worker updates.
arXiv Detail & Related papers (2024-02-05T12:26:01Z) - Byzantine Robustness and Partial Participation Can Be Achieved at Once: Just Clip Gradient Differences [61.74021364776313]
Distributed learning has emerged as a leading paradigm for training large machine learning models.
In real-world scenarios, participants may be unreliable or malicious, posing a significant challenge to the integrity and accuracy of the trained models.
We propose the first distributed method with client sampling and provable tolerance to Byzantine workers.
arXiv Detail & Related papers (2023-11-23T17:50:30Z) - Byzantine-Robust Loopless Stochastic Variance-Reduced Gradient [0.0]
We propose a new method -- Byzantine-Robust Loopless Variance Reduced Gradient (BR-LSVRG)
We derive non-asymptotic convergence guarantees for the new method in the strongly convex case.
arXiv Detail & Related papers (2023-03-08T13:20:49Z) - Byzantine-Robust Online and Offline Distributed Reinforcement Learning [60.970950468309056]
We consider a distributed reinforcement learning setting where multiple agents explore the environment and communicate their experiences through a central server.
$alpha$-fraction of agents are adversarial and can report arbitrary fake information.
We seek to identify a near-optimal policy for the underlying Markov decision process in the presence of these adversarial agents.
arXiv Detail & Related papers (2022-06-01T00:44:53Z) - Stochastic Alternating Direction Method of Multipliers for
Byzantine-Robust Distributed Learning [22.835940007753376]
We propose a Byzantine-robust alternating direction method of multipliers (ADMM) that fully utilizes the separable problem structure.
Theoretically, we prove that the proposed method converges to a bounded neighborhood of the optimal solution at a rate of O(k) under mild assumptions.
Numerical experiments on the MNIST and COVERTYPE datasets demonstrate the effectiveness of the proposed method to various Byzantine attacks.
arXiv Detail & Related papers (2021-06-13T01:17:31Z) - Learning from History for Byzantine Robust Optimization [52.68913869776858]
Byzantine robustness has received significant attention recently given its importance for distributed learning.
We show that most existing robust aggregation rules may not converge even in the absence of any Byzantine attackers.
arXiv Detail & Related papers (2020-12-18T16:22:32Z) - Byzantine-Robust Variance-Reduced Federated Learning over Distributed
Non-i.i.d. Data [36.99547890386817]
We consider the federated learning problem where data on workers are not independent and identically distributed (i.i.d.)
An unknown number of Byzantine workers may send malicious messages to the central node, leading to remarkable learning error.
Most of the Byzantine-robust methods address this issue by using robust aggregation rules to aggregate the received messages.
arXiv Detail & Related papers (2020-09-17T09:09:23Z) - Byzantine-Robust Learning on Heterogeneous Datasets via Bucketing [55.012801269326594]
In Byzantine robust distributed learning, a central server wants to train a machine learning model over data distributed across multiple workers.
A fraction of these workers may deviate from the prescribed algorithm and send arbitrary messages.
We propose a simple bucketing scheme that adapts existing robust algorithms to heterogeneous datasets at a negligible computational cost.
arXiv Detail & Related papers (2020-06-16T17:58:53Z) - Byzantine-resilient Decentralized Stochastic Gradient Descent [85.15773446094576]
We present an in-depth study towards the Byzantine resilience of decentralized learning systems.
We propose UBAR, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance.
arXiv Detail & Related papers (2020-02-20T05:11:04Z) - Federated Variance-Reduced Stochastic Gradient Descent with Robustness
to Byzantine Attacks [74.36161581953658]
This paper deals with distributed finite-sum optimization for learning over networks in the presence of malicious Byzantine attacks.
To cope with such attacks, most resilient approaches so far combine gradient descent (SGD) with different robust aggregation rules.
The present work puts forth a Byzantine attack resilient distributed (Byrd-) SAGA approach for learning tasks involving finite-sum optimization over networks.
arXiv Detail & Related papers (2019-12-29T19:46:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.