FPGA-Based Hardware Accelerator of Homomorphic Encryption for Efficient
Federated Learning
- URL: http://arxiv.org/abs/2007.10560v1
- Date: Tue, 21 Jul 2020 01:59:58 GMT
- Title: FPGA-Based Hardware Accelerator of Homomorphic Encryption for Efficient
Federated Learning
- Authors: Zhaoxiong Yang, Shuihai Hu, Kai Chen
- Abstract summary: Federated learning tends to utilize various privacy preserving mechanisms to protect the transferred intermediate data.
Maintaining accuracy and security more efficiently has been a key problem of federated learning.
Our framework implements the representative Paillier homomorphic cryptosystem with high level synthesis for flexibility and portability.
- Score: 9.733675923979108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the increasing awareness of privacy protection and data fragmentation
problem, federated learning has been emerging as a new paradigm of machine
learning. Federated learning tends to utilize various privacy preserving
mechanisms to protect the transferred intermediate data, among which
homomorphic encryption strikes a balance between security and ease of
utilization. However, the complicated operations and large operands impose
significant overhead on federated learning. Maintaining accuracy and security
more efficiently has been a key problem of federated learning. In this work, we
investigate a hardware solution, and design an FPGA-based homomorphic
encryption framework, aiming to accelerate the training phase in federated
learning. The root complexity lies in searching for a compact architecture for
the core operation of homomorphic encryption, to suit the requirement of
federated learning about high encryption throughput and flexibility of
configuration. Our framework implements the representative Paillier homomorphic
cryptosystem with high level synthesis for flexibility and portability, with
careful optimization on the modular multiplication operation in terms of
processing clock cycle, resource usage and clock frequency. Our accelerator
achieves a near-optimal execution clock cycle, with a better DSP-efficiency
than existing designs, and reduces the encryption time by up to 71% during
training process of various federated learning models.
Related papers
- QuanCrypt-FL: Quantized Homomorphic Encryption with Pruning for Secure Federated Learning [0.48342038441006796]
We propose QuanCrypt-FL, a novel algorithm that combines low-bit quantization and pruning techniques to enhance protection against attacks.
We validate our approach on MNIST, CIFAR-10, and CIFAR-100 datasets, demonstrating superior performance compared to state-of-the-art methods.
QuanCrypt-FL achieves up to 9x faster encryption, 16x faster decryption, and 1.5x faster inference compared to BatchCrypt, with training time reduced by up to 3x.
arXiv Detail & Related papers (2024-11-08T01:46:00Z) - Digital Twin-Assisted Federated Learning with Blockchain in Multi-tier Computing Systems [67.14406100332671]
In Industry 4.0 systems, resource-constrained edge devices engage in frequent data interactions.
This paper proposes a digital twin (DT) and federated digital twin (FL) scheme.
The efficacy of our proposed cooperative interference-based FL process has been verified through numerical analysis.
arXiv Detail & Related papers (2024-11-04T17:48:02Z) - Enhancing Spectrum Efficiency in 6G Satellite Networks: A GAIL-Powered Policy Learning via Asynchronous Federated Inverse Reinforcement Learning [67.95280175998792]
A novel adversarial imitation learning (GAIL)-powered policy learning approach is proposed for optimizing beamforming, spectrum allocation, and remote user equipment (RUE) association ins.
We employ inverse RL (IRL) to automatically learn reward functions without manual tuning.
We show that the proposed MA-AL method outperforms traditional RL approaches, achieving a $14.6%$ improvement in convergence and reward value.
arXiv Detail & Related papers (2024-09-27T13:05:02Z) - Federated Learning with Quantum Computing and Fully Homomorphic Encryption: A Novel Computing Paradigm Shift in Privacy-Preserving ML [4.92218040320554]
Federated Learning is a privacy-preserving alternative to conventional methods that allow multiple learning clients to share model knowledge without disclosing private data.
This work applies the Fully Homomorphic Encryption scheme to a Federated Learning Neural Network architecture that integrates both classical and quantum layers.
arXiv Detail & Related papers (2024-09-14T01:23:26Z) - Robust Representation Learning for Privacy-Preserving Machine Learning:
A Multi-Objective Autoencoder Approach [0.9831489366502302]
We propose a robust representation learning framework for privacy-preserving machine learning (ppML)
Our method centers on training autoencoders in a multi-objective manner and then concatenating the latent and learned features from the encoding part as the encoded form of our data.
With our proposed framework, we can share our data and use third party tools without being under the threat of revealing its original form.
arXiv Detail & Related papers (2023-09-08T16:41:25Z) - RoFL: Attestable Robustness for Secure Federated Learning [59.63865074749391]
Federated Learning allows a large number of clients to train a joint model without the need to share their private data.
To ensure the confidentiality of the client updates, Federated Learning systems employ secure aggregation.
We present RoFL, a secure Federated Learning system that improves robustness against malicious clients.
arXiv Detail & Related papers (2021-07-07T15:42:49Z) - Rethinking Architecture Design for Tackling Data Heterogeneity in
Federated Learning [53.73083199055093]
We show that attention-based architectures (e.g., Transformers) are fairly robust to distribution shifts.
Our experiments show that replacing convolutional networks with Transformers can greatly reduce catastrophic forgetting of previous devices.
arXiv Detail & Related papers (2021-06-10T21:04:18Z) - Efficient CNN Building Blocks for Encrypted Data [6.955451042536852]
Homomorphic Encryption (FHE) is a promising technique to enable machine learning and inferencing.
We show that operational parameters of the chosen FHE scheme have a major impact on the design of the machine learning model.
Our empirical study shows that choice of aforementioned design parameters result in significant trade-offs between accuracy, security level, and computational time.
arXiv Detail & Related papers (2021-01-30T21:47:23Z) - Faster Secure Data Mining via Distributed Homomorphic Encryption [108.77460689459247]
Homomorphic Encryption (HE) is receiving more and more attention recently for its capability to do computations over the encrypted field.
We propose a novel general distributed HE-based data mining framework towards one step of solving the scaling problem.
We verify the efficiency and effectiveness of our new framework by testing over various data mining algorithms and benchmark data-sets.
arXiv Detail & Related papers (2020-06-17T18:14:30Z) - Multilinear Compressive Learning with Prior Knowledge [106.12874293597754]
Multilinear Compressive Learning (MCL) framework combines Multilinear Compressive Sensing and Machine Learning into an end-to-end system.
Key idea behind MCL is the assumption of the existence of a tensor subspace which can capture the essential features from the signal for the downstream learning task.
In this paper, we propose a novel solution to address both of the aforementioned requirements, i.e., How to find those tensor subspaces in which the signals of interest are highly separable?
arXiv Detail & Related papers (2020-02-17T19:06:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.