Verifiable Coded Computing: Towards Fast, Secure and Private Distributed
Machine Learning
- URL: http://arxiv.org/abs/2107.12958v1
- Date: Tue, 27 Jul 2021 17:23:09 GMT
- Title: Verifiable Coded Computing: Towards Fast, Secure and Private Distributed
Machine Learning
- Authors: Tingting Tang, Ramy E. Ali, Hanieh Hashemi, Tynan Gangwani, Salman
Avestimehr and Murali Annavaram
- Abstract summary: Stragglers, Byzantine workers, and data privacy are the main bottlenecks in distributed cloud computing.
We propose Verifiable Coded Computing framework that decouples Byzantine node detection challenge from the straggler tolerance.
Our experiments show that VCC speeds up the conventional uncoded implementation of distributed logistic regression by $3.2times-6.9times$.
- Score: 13.09925205966904
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stragglers, Byzantine workers, and data privacy are the main bottlenecks in
distributed cloud computing. Several prior works proposed coded computing
strategies to jointly address all three challenges. They require either a large
number of workers, a significant communication cost or a significant
computational complexity to tolerate malicious workers. Much of the overhead in
prior schemes comes from the fact that they tightly couple coding for all three
problems into a single framework. In this work, we propose Verifiable Coded
Computing (VCC) framework that decouples Byzantine node detection challenge
from the straggler tolerance. VCC leverages coded computing just for handling
stragglers and privacy, and then uses an orthogonal approach of verifiable
computing to tackle Byzantine nodes. Furthermore, VCC dynamically adapts its
coding scheme to tradeoff straggler tolerance with Byzantine protection and
vice-versa. We evaluate VCC on compute intensive distributed logistic
regression application. Our experiments show that VCC speeds up the
conventional uncoded implementation of distributed logistic regression by
$3.2\times-6.9\times$, and also improves the test accuracy by up to $12.6\%$.
Related papers
- Decompiling Smart Contracts with a Large Language Model [51.49197239479266]
Despite Etherscan's 78,047,845 smart contracts deployed on (as of May 26, 2025), a mere 767,520 ( 1%) are open source.<n>This opacity necessitates the automated semantic analysis of on-chain smart contract bytecode.<n>We introduce a pioneering decompilation pipeline that transforms bytecode into human-readable and semantically faithful Solidity code.
arXiv Detail & Related papers (2025-06-24T13:42:59Z) - Fast correlated decoding of transversal logical algorithms [67.01652927671279]
Quantum error correction (QEC) is required for large-scale computation, but incurs a significant resource overhead.<n>Recent advances have shown that by jointly decoding logical qubits in algorithms composed of logical gates, the number of syndrome extraction rounds can be reduced.<n>Here, we reform the problem of decoding circuits by directly decoding relevant logical operator products as they propagate through the circuit.
arXiv Detail & Related papers (2025-05-19T18:00:00Z) - ByzSecAgg: A Byzantine-Resistant Secure Aggregation Scheme for Federated
Learning Based on Coded Computing and Vector Commitment [90.60126724503662]
ByzSecAgg is an efficient secure aggregation scheme for federated learning.
ByzSecAgg is protected against Byzantine attacks and privacy leakages.
arXiv Detail & Related papers (2023-02-20T11:15:18Z) - Inception-Based Crowd Counting -- Being Fast while Remaining Accurate [3.274290296343038]
A new method, based on Inception-V3, is proposed to reduce the amount of computation.
Experiments show that ICC can at best reduce 85.3 percent calculations with 24.4 percent performance loss.
arXiv Detail & Related papers (2022-10-18T12:12:13Z) - Implementing Reinforcement Learning Datacenter Congestion Control in NVIDIA NICs [64.26714148634228]
congestion control (CC) algorithms become extremely difficult to design.
It is currently not possible to deploy AI models on network devices due to their limited computational capabilities.
We build a computationally-light solution based on a recent reinforcement learning CC algorithm.
arXiv Detail & Related papers (2022-07-05T20:42:24Z) - Generalized Lagrange Coded Computing: A Flexible Computation-Communication Tradeoff for Resilient, Secure, and Private Computation [11.951700263976777]
Generalized Lagrange Coded Computing (GLCC) codes are proposed to provide resiliency against stragglers who do not return results in time.
LCC codes include the state-of-the-art Lagrange Coded Computing (LCC) codes as a special case.
arXiv Detail & Related papers (2022-04-24T02:48:07Z) - Secure Bilevel Asynchronous Vertical Federated Learning with Backward
Updating [159.48259714642447]
Vertical scalable learning (VFL) attracts increasing attention due to the demands of multi-party collaborative modeling and concerns of privacy leakage.
We propose a novel bftextlevel parallel architecture (VF$bfB2$), under which three new algorithms, including VF$B2$, are proposed.
arXiv Detail & Related papers (2021-03-01T12:34:53Z) - Faster Secure Data Mining via Distributed Homomorphic Encryption [108.77460689459247]
Homomorphic Encryption (HE) is receiving more and more attention recently for its capability to do computations over the encrypted field.
We propose a novel general distributed HE-based data mining framework towards one step of solving the scaling problem.
We verify the efficiency and effectiveness of our new framework by testing over various data mining algorithms and benchmark data-sets.
arXiv Detail & Related papers (2020-06-17T18:14:30Z) - Byzantine-Robust Learning on Heterogeneous Datasets via Bucketing [55.012801269326594]
In Byzantine robust distributed learning, a central server wants to train a machine learning model over data distributed across multiple workers.
A fraction of these workers may deviate from the prescribed algorithm and send arbitrary messages.
We propose a simple bucketing scheme that adapts existing robust algorithms to heterogeneous datasets at a negligible computational cost.
arXiv Detail & Related papers (2020-06-16T17:58:53Z) - Communication-Efficient Gradient Coding for Straggler Mitigation in
Distributed Learning [17.454251607446555]
Distributed implementations of gradient-based methods, wherein a server distributes gradient computations across worker machines, need to overcome two limitations.
Ye and Abbe [ICML 2018] proposed a coding-theoretic paradigm to characterize a fundamental trade-off between computation load per worker, communication overhead per worker, and straggler tolerance.
We develop a communication-efficient gradient coding framework to overcome these drawbacks.
arXiv Detail & Related papers (2020-05-14T17:57:13Z) - Straggler-aware Distributed Learning: Communication Computation Latency
Trade-off [56.08535873173518]
Straggling workers can be tolerated by assigning redundant computations and coding across data and computations.
In most existing schemes, each non-straggling worker transmits one message per iteration to the parameter server (PS) after completing all its computations.
Imposing such a limitation results in two main drawbacks; over-computation due to inaccurate prediction of the straggling behaviour, and under-utilization due to treating workers as straggler/non-straggler.
arXiv Detail & Related papers (2020-04-10T08:39:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.