Training Large Scale Polynomial CNNs for E2E Inference over Homomorphic
Encryption
- URL: http://arxiv.org/abs/2304.14836v2
- Date: Sun, 11 Jun 2023 10:07:52 GMT
- Title: Training Large Scale Polynomial CNNs for E2E Inference over Homomorphic
Encryption
- Authors: Moran Baruch, Nir Drucker, Gilad Ezov, Yoav Goldberg, Eyal Kushnir,
Jenny Lerner, Omri Soceanu and Itamar Zimerman
- Abstract summary: Training large-scale CNNs that during inference can be run under Homomorphic Encryption (HE) is challenging.
We provide a novel training method for large CNNs such as ResNet-152 and ConvNeXt models.
- Score: 33.35896071292604
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Training large-scale CNNs that during inference can be run under Homomorphic
Encryption (HE) is challenging due to the need to use only polynomial
operations. This limits HE-based solutions adoption. We address this challenge
and pioneer in providing a novel training method for large polynomial CNNs such
as ResNet-152 and ConvNeXt models, and achieve promising accuracy on encrypted
samples on large-scale dataset such as ImageNet. Additionally, we provide
optimization insights regarding activation functions and skip-connection
latency impacts, enhancing HE-based evaluation efficiency. Finally, to
demonstrate the robustness of our method, we provide a polynomial adaptation of
the CLIP model for secure zero-shot prediction, unlocking unprecedented
capabilities at the intersection of HE and transfer learning.
Related papers
- Adversarial Training of Two-Layer Polynomial and ReLU Activation Networks via Convex Optimization [40.68266398473983]
Training neural networks which are robust to adversarial attacks remains an important problem in deep learning.
We reformulate the training problems for two-layer ReLU andfty$ activation networks as convex programs.
We demonstrate the practical utility of convex adversarial training on large-scale problems.
arXiv Detail & Related papers (2024-05-22T22:08:13Z) - Dynamic Semantic Compression for CNN Inference in Multi-access Edge
Computing: A Graph Reinforcement Learning-based Autoencoder [82.8833476520429]
We propose a novel semantic compression method, autoencoder-based CNN architecture (AECNN) for effective semantic extraction and compression in partial offloading.
In the semantic encoder, we introduce a feature compression module based on the channel attention mechanism in CNNs, to compress intermediate data by selecting the most informative features.
In the semantic decoder, we design a lightweight decoder to reconstruct the intermediate data through learning from the received compressed data to improve accuracy.
arXiv Detail & Related papers (2024-01-19T15:19:47Z) - Regularized PolyKervNets: Optimizing Expressiveness and Efficiency for
Private Inference in Deep Neural Networks [0.0]
We focus on PolyKervNets, a technique known for offering improved dynamic approximations in smaller networks.
Our primary objective is to empirically explore optimization-based training recipes to enhance the performance of PolyKervNets in larger networks.
arXiv Detail & Related papers (2023-12-23T11:37:18Z) - The Convex Landscape of Neural Networks: Characterizing Global Optima
and Stationary Points via Lasso Models [75.33431791218302]
Deep Neural Network Network (DNN) models are used for programming purposes.
In this paper we examine the use of convex neural recovery models.
We show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
We also show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
arXiv Detail & Related papers (2023-12-19T23:04:56Z) - Towards a Better Theoretical Understanding of Independent Subnetwork Training [56.24689348875711]
We take a closer theoretical look at Independent Subnetwork Training (IST)
IST is a recently proposed and highly effective technique for solving the aforementioned problems.
We identify fundamental differences between IST and alternative approaches, such as distributed methods with compressed communication.
arXiv Detail & Related papers (2023-06-28T18:14:22Z) - Faster Adaptive Federated Learning [84.38913517122619]
Federated learning has attracted increasing attention with the emergence of distributed data.
In this paper, we propose an efficient adaptive algorithm (i.e., FAFED) based on momentum-based variance reduced technique in cross-silo FL.
arXiv Detail & Related papers (2022-12-02T05:07:50Z) - Fighting COVID-19 in the Dark: Methodology for Improved Inference Using
Homomorphically Encrypted DNN [3.1959970303072396]
homomorphic encryption (HE) has been used as a method to enable analytics while addressing privacy concerns.
There are several challenges related to the use of HE, including size limitations and the lack of support for some operation types.
We propose a structured methodology to replace ReLU with a quadratic activation.
arXiv Detail & Related papers (2021-11-05T10:04:15Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Active Learning in CNNs via Expected Improvement Maximization [2.0305676256390934]
"Dropout-based IMprOvementS" (DEIMOS) is a flexible and computationally-efficient approach to active learning.
Our results demonstrate that DEIMOS outperforms several existing baselines across multiple regression and classification tasks.
arXiv Detail & Related papers (2020-11-27T22:06:52Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.