VerifBFL: Leveraging zk-SNARKs for A Verifiable Blockchained Federated Learning
- URL: http://arxiv.org/abs/2501.04319v1
- Date: Wed, 08 Jan 2025 07:32:54 GMT
- Title: VerifBFL: Leveraging zk-SNARKs for A Verifiable Blockchained Federated Learning
- Authors: Ahmed Ayoub Bellachia, Mouhamed Amine Bouchiha, Yacine Ghamri-Doudane, Mourad Rabah,
- Abstract summary: This work presents VerifBFL, a trustless, privacy-preserving, and verifiable federated learning framework.
VerifBFL ensures the verifiability of both local training and aggregation processes.
To protect training data from inference attacks, VerifBFL leverages differential privacy.
- Score: 2.90114256542208
- License:
- Abstract: Blockchain-based Federated Learning (FL) is an emerging decentralized machine learning paradigm that enables model training without relying on a central server. Although some BFL frameworks are considered privacy-preserving, they are still vulnerable to various attacks, including inference and model poisoning. Additionally, most of these solutions employ strong trust assumptions among all participating entities or introduce incentive mechanisms to encourage collaboration, making them susceptible to multiple security flaws. This work presents VerifBFL, a trustless, privacy-preserving, and verifiable federated learning framework that integrates blockchain technology and cryptographic protocols. By employing zero-knowledge Succinct Non-Interactive Argument of Knowledge (zk-SNARKs) and incrementally verifiable computation (IVC), VerifBFL ensures the verifiability of both local training and aggregation processes. The proofs of training and aggregation are verified on-chain, guaranteeing the integrity and auditability of each participant's contributions. To protect training data from inference attacks, VerifBFL leverages differential privacy. Finally, to demonstrate the efficiency of the proposed protocols, we built a proof of concept using emerging tools. The results show that generating proofs for local training and aggregation in VerifBFL takes less than 81s and 2s, respectively, while verifying them on-chain takes less than 0.6s.
Related papers
- Enhancing Trust and Privacy in Distributed Networks: A Comprehensive Survey on Blockchain-based Federated Learning [51.13534069758711]
Decentralized approaches like blockchain offer a compelling solution by implementing a consensus mechanism among multiple entities.
Federated Learning (FL) enables participants to collaboratively train models while safeguarding data privacy.
This paper investigates the synergy between blockchain's security features and FL's privacy-preserving model training capabilities.
arXiv Detail & Related papers (2024-03-28T07:08:26Z) - PPBFL: A Privacy Protected Blockchain-based Federated Learning Model [6.278098707317501]
We propose a Protected-based Federated Learning Model (PPBFL) to enhance the security of federated learning.
We introduce a Proof of Training Work (PoTW) algorithm tailored for federated learning, aiming to incentive training nodes.
We also propose a new mix transactions mechanism utilizing ring signature technology to better protect the identity privacy of local training clients.
arXiv Detail & Related papers (2024-01-02T13:13:28Z) - zkDFL: An efficient and privacy-preserving decentralized federated
learning with zero-knowledge proof [3.517233208696287]
Federated learning (FL) has been widely adopted in various fields of study and business.
Traditional centralized FL systems suffer from serious issues.
We propose a zero-knowledge proof (ZKP)-based aggregator (zkDFL)
arXiv Detail & Related papers (2023-12-01T17:00:30Z) - FedChain: An Efficient and Secure Consensus Protocol based on Proof of Useful Federated Learning for Blockchain [0.3480973072524161]
The core of the blockchain is the consensus protocol, which establishes consensus among all the participants.
We propose an efficient and secure consensus protocol based on proof of useful federated learning for blockchain (called FedChain)
Our approach has been tested and validated through extensive experiments, demonstrating its performance.
arXiv Detail & Related papers (2023-08-29T08:04:07Z) - RoFL: Attestable Robustness for Secure Federated Learning [59.63865074749391]
Federated Learning allows a large number of clients to train a joint model without the need to share their private data.
To ensure the confidentiality of the client updates, Federated Learning systems employ secure aggregation.
We present RoFL, a secure Federated Learning system that improves robustness against malicious clients.
arXiv Detail & Related papers (2021-07-07T15:42:49Z) - CRFL: Certifiably Robust Federated Learning against Backdoor Attacks [59.61565692464579]
This paper provides the first general framework, Certifiably Robust Federated Learning (CRFL), to train certifiably robust FL models against backdoors.
Our method exploits clipping and smoothing on model parameters to control the global model smoothness, which yields a sample-wise robustness certification on backdoors with limited magnitude.
arXiv Detail & Related papers (2021-06-15T16:50:54Z) - Blockchain Assisted Decentralized Federated Learning (BLADE-FL):
Performance Analysis and Resource Allocation [119.19061102064497]
We propose a decentralized FL framework by integrating blockchain into FL, namely, blockchain assisted decentralized federated learning (BLADE-FL)
In a round of the proposed BLADE-FL, each client broadcasts its trained model to other clients, competes to generate a block based on the received models, and then aggregates the models from the generated block before its local training of the next round.
We explore the impact of lazy clients on the learning performance of BLADE-FL, and characterize the relationship among the optimal K, the learning parameters, and the proportion of lazy clients.
arXiv Detail & Related papers (2021-01-18T07:19:08Z) - Blockchain Assisted Decentralized Federated Learning (BLADE-FL) with
Lazy Clients [124.48732110742623]
We propose a novel framework by integrating blockchain into Federated Learning (FL)
BLADE-FL has a good performance in terms of privacy preservation, tamper resistance, and effective cooperation of learning.
It gives rise to a new problem of training deficiency, caused by lazy clients who plagiarize others' trained models and add artificial noises to conceal their cheating behaviors.
arXiv Detail & Related papers (2020-12-02T12:18:27Z) - Resource Management for Blockchain-enabled Federated Learning: A Deep
Reinforcement Learning Approach [54.29213445674221]
Federated Learning (BFL) enables mobile devices to collaboratively train neural network models required by a Machine Learning Model Owner (MLMO)
The issue of BFL is that the mobile devices have energy and CPU constraints that may reduce the system lifetime and training efficiency.
We propose to use the Deep Reinforcement Learning (DRL) to derive the optimal decisions for theO.
arXiv Detail & Related papers (2020-04-08T16:29:19Z) - A Blockchain-based Decentralized Federated Learning Framework with
Committee Consensus [20.787163387487816]
In mobile computing scenarios, federated learning protects users from exposing their private data, while cooperatively training the global model for a variety of real-world applications.
Security of federated learning is increasingly being questioned, due to the malicious clients or central servers' constant attack to the global model or user privacy data.
We propose a decentralized federated learning framework based on blockchain, i.e., a Committee consensus (BFLC) framework.
arXiv Detail & Related papers (2020-04-02T02:04:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.