Faster Secure Data Mining via Distributed Homomorphic Encryption
- URL: http://arxiv.org/abs/2006.10091v1
- Date: Wed, 17 Jun 2020 18:14:30 GMT
- Title: Faster Secure Data Mining via Distributed Homomorphic Encryption
- Authors: Junyi Li, Heng Huang
- Abstract summary: Homomorphic Encryption (HE) is receiving more and more attention recently for its capability to do computations over the encrypted field.
We propose a novel general distributed HE-based data mining framework towards one step of solving the scaling problem.
We verify the efficiency and effectiveness of our new framework by testing over various data mining algorithms and benchmark data-sets.
- Score: 108.77460689459247
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the rising privacy demand in data mining, Homomorphic Encryption (HE)
is receiving more and more attention recently for its capability to do
computations over the encrypted field. By using the HE technique, it is
possible to securely outsource model learning to the not fully trustful but
powerful public cloud computing environments. However, HE-based training scales
badly because of the high computation complexity. It is still an open problem
whether it is possible to apply HE to large-scale problems. In this paper, we
propose a novel general distributed HE-based data mining framework towards one
step of solving the scaling problem. The main idea of our approach is to use
the slightly more communication overhead in exchange of shallower computational
circuit in HE, so as to reduce the overall complexity. We verify the efficiency
and effectiveness of our new framework by testing over various data mining
algorithms and benchmark data-sets. For example, we successfully train a
logistic regression model to recognize the digit 3 and 8 within around 5
minutes, while a centralized counterpart needs almost 2 hours.
Related papers
- Optimal Oblivious Algorithms for Multi-way Joins [2.8151472703172398]
We propose a novel sorting-based algorithm for multi-way join processing that operates without relying on ORAM simulations or other security assumptions.
Our algorithm is a non-trivial, provably oblivious composition of basic primitives, with time complexity matching the insecure worst-case optimal join algorithm, up to a logarithmic factor.
arXiv Detail & Related papers (2025-01-08T01:23:29Z) - Evaluating the Potential of In-Memory Processing to Accelerate Homomorphic Encryption [1.5707609236065612]
homomorphic encryption (HE) allows computation without the need for decryption.
The high computational and memory overhead associated with the underlying cryptographic operations has hindered the practicality of HE-based solutions.
processing in-memory (PIM) presents a promising solution to this problem by bringing computation closer to data, thereby reducing the overhead resulting from processor-memory data movements.
arXiv Detail & Related papers (2024-12-12T10:28:58Z) - Hyperdimensional Computing as a Rescue for Efficient Privacy-Preserving
Machine Learning-as-a-Service [9.773163665697057]
Homomorphic encryption (HE) is a promising technique to address this adversity.
With HE, the service provider can take encrypted data as a query and run the model without decrypting it.
We show hyperdimensional computing can be a rescue for privacy-preserving machine learning over encrypted data.
arXiv Detail & Related papers (2023-08-17T00:25:17Z) - Solving Large-scale Spatial Problems with Convolutional Neural Networks [88.31876586547848]
We employ transfer learning to improve training efficiency for large-scale spatial problems.
We propose that a convolutional neural network (CNN) can be trained on small windows of signals, but evaluated on arbitrarily large signals with little to no performance degradation.
arXiv Detail & Related papers (2023-06-14T01:24:42Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - Communication-Efficient Adam-Type Algorithms for Distributed Data Mining [93.50424502011626]
We propose a class of novel distributed Adam-type algorithms (emphi.e., SketchedAMSGrad) utilizing sketching.
Our new algorithm achieves a fast convergence rate of $O(frac1sqrtnT + frac1(k/d)2 T)$ with the communication cost of $O(k log(d))$ at each iteration.
arXiv Detail & Related papers (2022-10-14T01:42:05Z) - CryptoGCN: Fast and Scalable Homomorphically Encrypted Graph
Convolutional Network Inference [12.03953896181613]
Cloud-based graph convolutional network (GCN) has demonstrated great success and potential in many privacy-sensitive applications.
Despite its high inference accuracy and performance on cloud, maintaining data privacy in GCN inference remains largely unexplored.
In this paper, we take an initial attempt towards this and develop $textitCryptoGCN$--a homomorphic encryption (HE) based GCN inference framework.
arXiv Detail & Related papers (2022-09-24T02:20:54Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - Byzantine-Robust Learning on Heterogeneous Datasets via Bucketing [55.012801269326594]
In Byzantine robust distributed learning, a central server wants to train a machine learning model over data distributed across multiple workers.
A fraction of these workers may deviate from the prescribed algorithm and send arbitrary messages.
We propose a simple bucketing scheme that adapts existing robust algorithms to heterogeneous datasets at a negligible computational cost.
arXiv Detail & Related papers (2020-06-16T17:58:53Z) - Privacy-Preserving Gaussian Process Regression -- A Modular Approach to
the Application of Homomorphic Encryption [4.1499725848998965]
Homomorphic encryption (FHE) allows data to be computed on whilst encrypted.
Some commonly used machine learning algorithms, such as Gaussian process regression, are poorly suited to FHE.
We show that a modular approach, which applies FHE to only the sensitive steps of a workflow that need protection, allows one party to make predictions on their data.
arXiv Detail & Related papers (2020-01-28T11:50:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.