SIRNN: A Math Library for Secure RNN Inference
- URL: http://arxiv.org/abs/2105.04236v1
- Date: Mon, 10 May 2021 10:04:46 GMT
- Title: SIRNN: A Math Library for Secure RNN Inference
- Authors: Deevashwer Rathee, Mayank Rathee, Rahul Kranti Kiran Goli, Divya
Gupta, Rahul Sharma, Nishanth Chandran, Aseem Rastogi
- Abstract summary: We provide new specialized 2PC protocols for math functions that crucially rely on lookup-tables and mixed-bitwidths.
Our protocols for math functions communicate up to 423x less data than prior work.
We build on top of our novel protocols to build SIRNN, a library for end-to-end secure 2-party inference.
- Score: 6.323743920987275
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Complex machine learning (ML) inference algorithms like recurrent neural
networks (RNNs) use standard functions from math libraries like exponentiation,
sigmoid, tanh, and reciprocal of square root. Although prior work on secure
2-party inference provides specialized protocols for convolutional neural
networks (CNNs), existing secure implementations of these math operators rely
on generic 2-party computation (2PC) protocols that suffer from high
communication. We provide new specialized 2PC protocols for math functions that
crucially rely on lookup-tables and mixed-bitwidths to address this performance
overhead; our protocols for math functions communicate up to 423x less data
than prior work. Some of the mixed bitwidth operations used by our math
implementations are (zero and signed) extensions, different forms of
truncations, multiplication of operands of mixed-bitwidths, and digit
decomposition (a generalization of bit decomposition to larger digits). For
each of these primitive operations, we construct specialized 2PC protocols that
are more communication efficient than generic 2PC, and can be of independent
interest. Furthermore, our math implementations are numerically precise, which
ensures that the secure implementations preserve model accuracy of cleartext.
We build on top of our novel protocols to build SIRNN, a library for end-to-end
secure 2-party DNN inference, that provides the first secure implementations of
an RNN operating on time series sensor data, an RNN operating on speech data,
and a state-of-the-art ML architecture that combines CNNs and RNNs for
identifying all heads present in images. Our evaluation shows that SIRNN
achieves up to three orders of magnitude of performance improvement when
compared to inference of these models using an existing state-of-the-art 2PC
framework.
Related papers
- Enhancing MOTION2NX for Efficient, Scalable and Secure Image Inference using Convolutional Neural Networks [4.407841002228536]
We use the ABY2.0 SMPC protocol implemented on the C++ based MOTION2NX framework for secure convolutional neural network (CNN) inference application with semi-honest security.
We also present a novel splitting algorithm that divides the computations at each CNN layer into multiple chunks.
arXiv Detail & Related papers (2024-08-29T09:50:21Z) - OPAF: Optimized Secure Two-Party Computation Protocols for Nonlinear Activation Functions in Recurrent Neural Network [8.825150825838769]
This paper pays special attention to the implementation of non-linear functions in semi-honest model with two-party settings.
We propose a novel and efficient protocol for exponential function by using a divide-and-conquer strategy.
Next, we take advantage of the symmetry of sigmoid and Tanh, and fine-tune the inputs to reduce the 2PC building blocks.
arXiv Detail & Related papers (2024-03-01T02:49:40Z) - SVNet: Where SO(3) Equivariance Meets Binarization on Point Cloud
Representation [65.4396959244269]
The paper tackles the challenge by designing a general framework to construct 3D learning architectures.
The proposed approach can be applied to general backbones like PointNet and DGCNN.
Experiments on ModelNet40, ShapeNet, and the real-world dataset ScanObjectNN, demonstrated that the method achieves a great trade-off between efficiency, rotation, and accuracy.
arXiv Detail & Related papers (2022-09-13T12:12:19Z) - Towards Accurate Binary Neural Networks via Modeling Contextual
Dependencies [52.691032025163175]
Existing Binary Neural Networks (BNNs) operate mainly on local convolutions with binarization function.
We present new designs of binary neural modules, which enables leading binary neural modules by a large margin.
arXiv Detail & Related papers (2022-09-03T11:51:04Z) - Ultra-compact Binary Neural Networks for Human Activity Recognition on
RISC-V Processors [10.195581493173643]
Human Activity Recognition (HAR) is a relevant inference task in many mobile applications.
We propose a novel implementation of HAR based on deep neural networks, and precisely on Binary Neural Networks (BNNs)
BNNs yield very small memory footprints and low inference complexity, thanks to the replacement of arithmetic operations with bit-wise ones.
arXiv Detail & Related papers (2022-05-25T13:52:35Z) - iRNN: Integer-only Recurrent Neural Network [0.8766022970635899]
We present a quantization-aware training method for obtaining a highly accurate integer-only recurrent neural network (iRNN)
Our iRNN maintains similar performance as its full-precision counterpart, their deployment on smartphones improves the runtime performance by $2times$, and reduces the model size by $4times$.
arXiv Detail & Related papers (2021-09-20T20:17:40Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Solving Mixed Integer Programs Using Neural Networks [57.683491412480635]
This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one.
Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP.
We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each.
arXiv Detail & Related papers (2020-12-23T09:33:11Z) - CrypTFlow2: Practical 2-Party Secure Inference [10.733878947770283]
CrypTFlow2 is a cryptographic framework for secure inference over realistic Deep Neural Networks (DNNs)
We present the first secure inference over ImageNet-scale DNNs like ResNet50 and DenseNet121.
CrypTFlow2 requires an order of magnitude less communication and 20x-30x less time than the state-of-the-art.
arXiv Detail & Related papers (2020-10-13T15:12:28Z) - FATNN: Fast and Accurate Ternary Neural Networks [89.07796377047619]
Ternary Neural Networks (TNNs) have received much attention due to being potentially orders of magnitude faster in inference, as well as more power efficient, than full-precision counterparts.
In this work, we show that, under some mild constraints, computational complexity of the ternary inner product can be reduced by a factor of 2.
We elaborately design an implementation-dependent ternary quantization algorithm to mitigate the performance gap.
arXiv Detail & Related papers (2020-08-12T04:26:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.