2CP: Decentralized Protocols to Transparently Evaluate Contributivity in
Blockchain Federated Learning Environments
- URL: http://arxiv.org/abs/2011.07516v1
- Date: Sun, 15 Nov 2020 12:59:56 GMT
- Title: 2CP: Decentralized Protocols to Transparently Evaluate Contributivity in
Blockchain Federated Learning Environments
- Authors: Harry Cai and Daniel Rueckert and Jonathan Passerat-Palmbach
- Abstract summary: We introduce 2CP, a framework comprising two novel protocols for Federated Learning.
Crowdsource Protocol allows an actor to bring a model forward for training, and use their own data to evaluate the contributions made to it.
The Consortium Protocol gives trainers the same guarantee even when no party owns the initial model and no dataset is available.
- Score: 9.885896204530878
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Federated Learning harnesses data from multiple sources to build a single
model. While the initial model might belong solely to the actor bringing it to
the network for training, determining the ownership of the trained model
resulting from Federated Learning remains an open question. In this paper we
explore how Blockchains (in particular Ethereum) can be used to determine the
evolving ownership of a model trained with Federated Learning.
Firstly, we use the step-by-step evaluation metric to assess the relative
contributivities of participants in a Federated Learning process. Next, we
introduce 2CP, a framework comprising two novel protocols for Blockchained
Federated Learning, which both reward contributors with shares in the final
model based on their relative contributivity. The Crowdsource Protocol allows
an actor to bring a model forward for training, and use their own data to
evaluate the contributions made to it. Potential trainers are guaranteed a fair
share of the resulting model, even in a trustless setting. The Consortium
Protocol gives trainers the same guarantee even when no party owns the initial
model and no evaluator is available.
We conduct experiments with the MNIST dataset that reveal sound
contributivity scores resulting from both Protocols by rewarding larger
datasets with greater shares in the model. Our experiments also showed the
necessity to pair 2CP with a robust model aggregation mechanism to discard low
quality inputs coming from model poisoning attacks.
Related papers
- Secrets of RLHF in Large Language Models Part II: Reward Modeling [134.97964938009588]
We introduce a series of novel methods to mitigate the influence of incorrect and ambiguous preferences in the dataset.
We also introduce contrastive learning to enhance the ability of reward models to distinguish between chosen and rejected responses.
arXiv Detail & Related papers (2024-01-11T17:56:59Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - A Secure Aggregation for Federated Learning on Long-Tailed Data [26.168909973264707]
Federated Learning (FL) faces two challenges: the unbalanced distribution of training data among participants, and the model attack by Byzantine nodes.
A novel two-layer aggregation method is proposed for the rejection of malicious models and the advisable selection of valuable models.
Preliminary experiments validate that the think tank can make effective model selections for global aggregation.
arXiv Detail & Related papers (2023-07-17T08:42:21Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - A Marketplace for Trading AI Models based on Blockchain and Incentives
for IoT Data [24.847898465750667]
An emerging paradigm in Machine Learning (ML) is a federated approach where the learning model is delivered to a group of heterogeneous agents partially, allowing agents to train the model locally with their own data.
The problem of valuation of models, as well as the questions of incentives for collaborative training and trading of data/models, have received limited treatment in the literature.
In this paper, a new ecosystem of ML model trading over a trusted ML-based network is proposed. The buyer can acquire the model of interest from the ML market, and interested sellers spend local computations on their data to enhance that model's quality
arXiv Detail & Related papers (2021-12-06T08:52:42Z) - Partner-Assisted Learning for Few-Shot Image Classification [54.66864961784989]
Few-shot Learning has been studied to mimic human visual capabilities and learn effective models without the need of exhaustive human annotation.
In this paper, we focus on the design of training strategy to obtain an elemental representation such that the prototype of each novel class can be estimated from a few labeled samples.
We propose a two-stage training scheme, which first trains a partner encoder to model pair-wise similarities and extract features serving as soft-anchors, and then trains a main encoder by aligning its outputs with soft-anchors while attempting to maximize classification performance.
arXiv Detail & Related papers (2021-09-15T22:46:19Z) - Decentralized Federated Learning Preserves Model and Data Privacy [77.454688257702]
We propose a fully decentralized approach, which allows to share knowledge between trained models.
Students are trained on the output of their teachers via synthetically generated input data.
The results show that an untrained student model, trained on the teachers output reaches comparable F1-scores as the teacher.
arXiv Detail & Related papers (2021-02-01T14:38:54Z) - Transparent Contribution Evaluation for Secure Federated Learning on
Blockchain [10.920274650337559]
We propose a blockchain-based federated learning framework and a protocol to transparently evaluate each participants' contribution.
Our framework protects all parties' privacy in the model building phrase and transparently evaluates contributions based on the model updates.
arXiv Detail & Related papers (2021-01-26T05:49:59Z) - Blockchain Assisted Decentralized Federated Learning (BLADE-FL):
Performance Analysis and Resource Allocation [119.19061102064497]
We propose a decentralized FL framework by integrating blockchain into FL, namely, blockchain assisted decentralized federated learning (BLADE-FL)
In a round of the proposed BLADE-FL, each client broadcasts its trained model to other clients, competes to generate a block based on the received models, and then aggregates the models from the generated block before its local training of the next round.
We explore the impact of lazy clients on the learning performance of BLADE-FL, and characterize the relationship among the optimal K, the learning parameters, and the proportion of lazy clients.
arXiv Detail & Related papers (2021-01-18T07:19:08Z) - Analysis of Models for Decentralized and Collaborative AI on Blockchain [0.0]
We evaluate the use of several models and configurations in order to propose best practices when using the Self-Assessment incentive mechanism.
We compare several factors for each dataset when models are hosted in smart contracts on a public blockchain.
arXiv Detail & Related papers (2020-09-14T21:38:55Z) - BlockFLow: An Accountable and Privacy-Preserving Solution for Federated
Learning [2.0625936401496237]
BlockFLow is an accountable federated learning system that is fully decentralized and privacy-preserving.
Its primary goal is to reward agents proportional to the quality of their contribution while protecting the privacy of the underlying datasets and being resilient to malicious adversaries.
arXiv Detail & Related papers (2020-07-08T02:24:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.