Faster Secure Data Mining via Distributed Homomorphic Encryption
- URL: http://arxiv.org/abs/2006.10091v1
- Date: Wed, 17 Jun 2020 18:14:30 GMT
- Title: Faster Secure Data Mining via Distributed Homomorphic Encryption
- Authors: Junyi Li, Heng Huang
- Abstract summary: Homomorphic Encryption (HE) is receiving more and more attention recently for its capability to do computations over the encrypted field.
We propose a novel general distributed HE-based data mining framework towards one step of solving the scaling problem.
We verify the efficiency and effectiveness of our new framework by testing over various data mining algorithms and benchmark data-sets.
- Score: 108.77460689459247
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the rising privacy demand in data mining, Homomorphic Encryption (HE)
is receiving more and more attention recently for its capability to do
computations over the encrypted field. By using the HE technique, it is
possible to securely outsource model learning to the not fully trustful but
powerful public cloud computing environments. However, HE-based training scales
badly because of the high computation complexity. It is still an open problem
whether it is possible to apply HE to large-scale problems. In this paper, we
propose a novel general distributed HE-based data mining framework towards one
step of solving the scaling problem. The main idea of our approach is to use
the slightly more communication overhead in exchange of shallower computational
circuit in HE, so as to reduce the overall complexity. We verify the efficiency
and effectiveness of our new framework by testing over various data mining
algorithms and benchmark data-sets. For example, we successfully train a
logistic regression model to recognize the digit 3 and 8 within around 5
minutes, while a centralized counterpart needs almost 2 hours.
Related papers
- A Novel Neural Network-Based Federated Learning System for Imbalanced
and Non-IID Data [2.9642661320713555]
Most machine learning algorithms rely heavily on large amount of data which may be collected from various sources.
To combat this issue, researchers have introduced federated learning, where a prediction model is learnt by ensuring the privacy of data of clients data.
In this research, we propose a centralized, neural network-based federated learning system.
arXiv Detail & Related papers (2023-11-16T17:14:07Z) - Hyperdimensional Computing as a Rescue for Efficient Privacy-Preserving
Machine Learning-as-a-Service [9.773163665697057]
Homomorphic encryption (HE) is a promising technique to address this adversity.
With HE, the service provider can take encrypted data as a query and run the model without decrypting it.
We show hyperdimensional computing can be a rescue for privacy-preserving machine learning over encrypted data.
arXiv Detail & Related papers (2023-08-17T00:25:17Z) - Solving Large-scale Spatial Problems with Convolutional Neural Networks [88.31876586547848]
We employ transfer learning to improve training efficiency for large-scale spatial problems.
We propose that a convolutional neural network (CNN) can be trained on small windows of signals, but evaluated on arbitrarily large signals with little to no performance degradation.
arXiv Detail & Related papers (2023-06-14T01:24:42Z) - Dash: Accelerating Distributed Private Convolutional Neural Network Inference with Arithmetic Garbled Circuits [6.912820984005411]
We present Dash, a fast and distributed private convolutional neural network inference scheme secure against malicious attackers.
Building on arithmetic garbling gadgets [BMR16] and fancy-garbling [BCM+19], Dash is based purely on arithmetic garbled circuits.
arXiv Detail & Related papers (2023-02-13T13:48:08Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - Communication-Efficient Adam-Type Algorithms for Distributed Data Mining [93.50424502011626]
We propose a class of novel distributed Adam-type algorithms (emphi.e., SketchedAMSGrad) utilizing sketching.
Our new algorithm achieves a fast convergence rate of $O(frac1sqrtnT + frac1(k/d)2 T)$ with the communication cost of $O(k log(d))$ at each iteration.
arXiv Detail & Related papers (2022-10-14T01:42:05Z) - CryptoGCN: Fast and Scalable Homomorphically Encrypted Graph
Convolutional Network Inference [12.03953896181613]
Cloud-based graph convolutional network (GCN) has demonstrated great success and potential in many privacy-sensitive applications.
Despite its high inference accuracy and performance on cloud, maintaining data privacy in GCN inference remains largely unexplored.
In this paper, we take an initial attempt towards this and develop $textitCryptoGCN$--a homomorphic encryption (HE) based GCN inference framework.
arXiv Detail & Related papers (2022-09-24T02:20:54Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - Byzantine-Robust Learning on Heterogeneous Datasets via Bucketing [55.012801269326594]
In Byzantine robust distributed learning, a central server wants to train a machine learning model over data distributed across multiple workers.
A fraction of these workers may deviate from the prescribed algorithm and send arbitrary messages.
We propose a simple bucketing scheme that adapts existing robust algorithms to heterogeneous datasets at a negligible computational cost.
arXiv Detail & Related papers (2020-06-16T17:58:53Z) - Privacy-Preserving Gaussian Process Regression -- A Modular Approach to
the Application of Homomorphic Encryption [4.1499725848998965]
Homomorphic encryption (FHE) allows data to be computed on whilst encrypted.
Some commonly used machine learning algorithms, such as Gaussian process regression, are poorly suited to FHE.
We show that a modular approach, which applies FHE to only the sensitive steps of a workflow that need protection, allows one party to make predictions on their data.
arXiv Detail & Related papers (2020-01-28T11:50:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.