MPC-Pipe: an Efficient Pipeline Scheme for Secure Multi-party Machine
Learning Inference
- URL: http://arxiv.org/abs/2209.13643v1
- Date: Tue, 27 Sep 2022 19:16:26 GMT
- Title: MPC-Pipe: an Efficient Pipeline Scheme for Secure Multi-party Machine
Learning Inference
- Authors: Yongqin Wang, Rachit Rajat, Murali Annavaram
- Abstract summary: Multi-party computing (MPC) has been gaining popularity over the past years as a secure computing model.
MPC has fewer overheads than homomorphic encryption (HE) and has a more robust threat model than hardware-based trusted execution environments.
MPC protocols still pay substantial performance penalties compared to plaintext when applied to machine learning algorithms.
- Score: 3.1853566662905943
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Multi-party computing (MPC) has been gaining popularity over the past years
as a secure computing model, particularly for machine learning (ML) inference.
Compared with its competitors, MPC has fewer overheads than homomorphic
encryption (HE) and has a more robust threat model than hardware-based trusted
execution environments (TEE) such as Intel SGX. Despite its apparent
advantages, MPC protocols still pay substantial performance penalties compared
to plaintext when applied to ML algorithms. The overhead is due to added
computation and communication costs. For multiplications that are ubiquitous in
ML algorithms, MPC protocols add 32x more computational costs and 1 round of
broadcasting among MPC servers. Moreover, ML computations that have trivial
costs in plaintext, such as Softmax, ReLU, and other non-linear operations
become very expensive due to added communication. Those added overheads make
MPC less palatable to deploy in real-time ML inference frameworks, such as
speech translation.
In this work, we present MPC-Pipe, an MPC pipeline inference technique that
uses two ML-specific approaches. 1) inter-linear-layer pipeline and 2) inner
layer pipeline. Those two techniques shorten the total inference runtime for
machine learning models. Our experiments have shown to reduce ML inference
latency by up to 12.6% when model weights are private and 14.48\% when model
weights are public, compared to current MPC protocol implementations.
Related papers
- Efficient Private Inference Based on Helper-Assisted Malicious Security Dishonest Majority MPC [5.797285315996385]
We propose a novel, three-layer private inference framework based on the Helper-Assisted MSDM model.<n>The framework achieves up to a 2.4-25.7x speedup in LAN and a 1.3-9.5x acceleration in WAN over the state-of-the-art MSDM frameworks.
arXiv Detail & Related papers (2025-07-13T12:24:02Z) - MPCritic: A plug-and-play MPC architecture for reinforcement learning [6.656737591902601]
This paper presents MPCritic, a machine learning-friendly architecture that interfaces seamlessly with MPC tools.
MPCritic utilizes the loss landscape defined by a parameterized MPC problem, focusing on "soft" optimization over batched training steps.
arXiv Detail & Related papers (2025-04-01T18:07:07Z) - DMM: Distributed Matrix Mechanism for Differentially-Private Federated Learning Based on Constant-Overhead Linear Secret Resharing [51.336015600778396]
We introduce the distributed matrix mechanism to achieve the best-of-both-worlds; better privacy of distributed DP and better utility from the matrix mechanism.<n>We accomplish this using a novel cryptographic protocol that securely transfers sensitive values across client committees of different training iterations with constant communication overhead.
arXiv Detail & Related papers (2024-10-21T16:25:14Z) - Low-Latency Privacy-Preserving Deep Learning Design via Secure MPC [31.35072624488929]
Secure multi-party computation (MPC) facilitates privacy-preserving computation between multiple parties without leaking private information.
This work proposes a low-latency secret-sharing-based MPC design that reduces unnecessary communication rounds during the execution of MPC protocols.
arXiv Detail & Related papers (2024-07-24T07:01:21Z) - Parameter-Adaptive Approximate MPC: Tuning Neural-Network Controllers without Retraining [50.00291020618743]
This work introduces a novel, parameter-adaptive AMPC architecture capable of online tuning without recomputing large datasets and retraining.
We showcase the effectiveness of parameter-adaptive AMPC by controlling the swing-ups of two different real cartpole systems with a severely resource-constrained microcontroller (MCU)
Taken together, these contributions represent a marked step toward the practical application of AMPC in real-world systems.
arXiv Detail & Related papers (2024-04-08T20:02:19Z) - On Building Myopic MPC Policies using Supervised Learning [0.0]
This paper considers an alternative strategy, where supervised learning is used to learn the optimal value function offline instead of learning the optimal policy.
This can then be used as the cost-to-go function in a myopic MPC with a very short prediction horizon.
arXiv Detail & Related papers (2024-01-23T08:08:09Z) - Fast and Private Inference of Deep Neural Networks by Co-designing Activation Functions [26.125340303868335]
Current approaches suffer from large inference times.
We propose a novel training algorithm that gives accuracy competitive with inferences models.
Our evaluation shows between $3$ and $110times$ speedups in inference time on large models with up to $23$ million parameters.
arXiv Detail & Related papers (2023-06-14T14:38:25Z) - Accelerating Wireless Federated Learning via Nesterov's Momentum and
Distributed Principle Component Analysis [59.127630388320036]
A wireless federated learning system is investigated by allowing a server and workers to exchange uncoded information via wireless channels.
Since the workers frequently upload local to the server via bandwidth-limited channels, the uplink transmission from the workers to the server becomes a communication bottleneck.
A one-shot distributed principle component analysis (PCA) is leveraged to reduce the dimension of the dimension of the communication bottleneck.
arXiv Detail & Related papers (2023-03-31T08:41:42Z) - Partitioning Distributed Compute Jobs with Reinforcement Learning and
Graph Neural Networks [58.720142291102135]
Large-scale machine learning models are bringing advances to a broad range of fields.
Many of these models are too large to be trained on a single machine, and must be distributed across multiple devices.
We show that maximum parallelisation is sub-optimal in relation to user-critical metrics such as throughput and blocking rate.
arXiv Detail & Related papers (2023-01-31T17:41:07Z) - Fair and Efficient Distributed Edge Learning with Hybrid Multipath TCP [62.81300791178381]
The bottleneck of distributed edge learning over wireless has shifted from computing to communication.
Existing TCP-based data networking schemes for DEL are application-agnostic and fail to deliver adjustments according to application layer requirements.
We develop a hybrid multipath TCP (MP TCP) by combining model-based and deep reinforcement learning (DRL) based MP TCP for DEL.
arXiv Detail & Related papers (2022-11-03T09:08:30Z) - Towards Semantic Communication Protocols: A Probabilistic Logic
Perspective [69.68769942563812]
We propose a semantic protocol model (SPM) constructed by transforming an NPM into an interpretable symbolic graph written in the probabilistic logic programming language (ProbLog)
By leveraging its interpretability and memory-efficiency, we demonstrate several applications such as SPM reconfiguration for collision-avoidance.
arXiv Detail & Related papers (2022-07-08T14:19:36Z) - HD-cos Networks: Efficient Neural Architectures for Secure Multi-Party
Computation [26.67099154998755]
Multi-party computation (MPC) is a branch of cryptography where multiple non-colluding parties execute a protocol to securely compute a function.
We study training and inference of neural networks under the MPC setup.
We show that both of the approaches enjoy strong theoretical motivations and efficient computation under the MPC setup.
arXiv Detail & Related papers (2021-10-28T21:15:11Z) - Adam in Private: Secure and Fast Training of Deep Neural Networks with
Adaptive Moment Estimation [6.342794803074475]
We propose a framework that allows efficient evaluation of full-fledged state-of-the-art machine learning algorithms.
This is in contrast to most prior works, which substitute ML algorithms with approximated "MPC-friendly" variants.
We obtain secure training that outperforms state-of-the-art three-party systems.
arXiv Detail & Related papers (2021-06-04T01:40:09Z) - Einsum Networks: Fast and Scalable Learning of Tractable Probabilistic
Circuits [99.59941892183454]
We propose Einsum Networks (EiNets), a novel implementation design for PCs.
At their core, EiNets combine a large number of arithmetic operations in a single monolithic einsum-operation.
We show that the implementation of Expectation-Maximization (EM) can be simplified for PCs, by leveraging automatic differentiation.
arXiv Detail & Related papers (2020-04-13T23:09:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.