Secure Distributed Training at Scale
- URL: http://arxiv.org/abs/2106.11257v1
- Date: Mon, 21 Jun 2021 17:00:42 GMT
- Title: Secure Distributed Training at Scale
- Authors: Eduard Gorbunov, Alexander Borzunov, Michael Diskin, Max Ryabinin
- Abstract summary: Training in presence of peers requires specialized distributed training algorithms with Byzantine tolerance.
We propose a novel protocol for secure (Byzantine-tolerant) decentralized training that emphasizes communication efficiency.
- Score: 65.7538150168154
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Some of the hardest problems in deep learning can be solved with the combined
effort of many independent parties, as is the case for volunteer computing and
federated learning. These setups rely on high numbers of peers to provide
computational resources or train on decentralized datasets. Unfortunately,
participants in such systems are not always reliable. Any single participant
can jeopardize the entire training run by sending incorrect updates, whether
deliberately or by mistake. Training in presence of such peers requires
specialized distributed training algorithms with Byzantine tolerance. These
algorithms often sacrifice efficiency by introducing redundant communication or
passing all updates through a trusted server. As a result, it can be infeasible
to apply such algorithms to large-scale distributed deep learning, where models
can have billions of parameters. In this work, we propose a novel protocol for
secure (Byzantine-tolerant) decentralized training that emphasizes
communication efficiency. We rigorously analyze this protocol: in particular,
we provide theoretical bounds for its resistance against Byzantine and Sybil
attacks and show that it has a marginal communication overhead. To demonstrate
its practical effectiveness, we conduct large-scale experiments on image
classification and language modeling in presence of Byzantine attackers.
Related papers
- Mitigating Data Injection Attacks on Federated Learning [20.24380409762923]
Federated learning is a technique that allows multiple entities to collaboratively train models using their data.
Despite its advantages, federated learning can be susceptible to false data injection attacks.
We propose a novel technique to detect and mitigate data injection attacks on federated learning systems.
arXiv Detail & Related papers (2023-12-04T18:26:31Z) - Communication-Efficient Decentralized Federated Learning via One-Bit
Compressive Sensing [52.402550431781805]
Decentralized federated learning (DFL) has gained popularity due to its practicality across various applications.
Compared to the centralized version, training a shared model among a large number of nodes in DFL is more challenging.
We develop a novel algorithm based on the framework of the inexact alternating direction method (iADM)
arXiv Detail & Related papers (2023-08-31T12:22:40Z) - Combating Exacerbated Heterogeneity for Robust Models in Federated
Learning [91.88122934924435]
Combination of adversarial training and federated learning can lead to the undesired robustness deterioration.
We propose a novel framework called Slack Federated Adversarial Training (SFAT)
We verify the rationality and effectiveness of SFAT on various benchmarked and real-world datasets.
arXiv Detail & Related papers (2023-03-01T06:16:15Z) - Distributed Adversarial Training to Robustify Deep Neural Networks at
Scale [100.19539096465101]
Current deep neural networks (DNNs) are vulnerable to adversarial attacks, where adversarial perturbations to the inputs can change or manipulate classification.
To defend against such attacks, an effective approach, known as adversarial training (AT), has been shown to mitigate robust training.
We propose a large-batch adversarial training framework implemented over multiple machines.
arXiv Detail & Related papers (2022-06-13T15:39:43Z) - Federated Learning via Plurality Vote [38.778944321534084]
Federated learning allows collaborative workers to solve a machine learning problem while preserving data privacy.
Recent studies have tackled various challenges in federated learning.
We propose a new scheme named federated learning via plurality vote (FedVote)
arXiv Detail & Related papers (2021-10-06T18:16:22Z) - Byzantine-Robust Learning on Heterogeneous Datasets via Bucketing [55.012801269326594]
In Byzantine robust distributed learning, a central server wants to train a machine learning model over data distributed across multiple workers.
A fraction of these workers may deviate from the prescribed algorithm and send arbitrary messages.
We propose a simple bucketing scheme that adapts existing robust algorithms to heterogeneous datasets at a negligible computational cost.
arXiv Detail & Related papers (2020-06-16T17:58:53Z) - Communication-Efficient Decentralized Learning with Sparsification and
Adaptive Peer Selection [13.963329236804586]
We introduce a novel decentralized training algorithm with the following key features.
Each worker only needs to communicate with a single peer at each communication round with a highly compressed model.
Experimental results show that our algorithm significantly reduces the communication traffic and generally selects relatively high bandwidth peers.
arXiv Detail & Related papers (2020-02-22T12:31:57Z) - Byzantine-resilient Decentralized Stochastic Gradient Descent [85.15773446094576]
We present an in-depth study towards the Byzantine resilience of decentralized learning systems.
We propose UBAR, a novel algorithm to enhance decentralized learning with Byzantine Fault Tolerance.
arXiv Detail & Related papers (2020-02-20T05:11:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.