GTG-Shapley: Efficient and Accurate Participant Contribution Evaluation
in Federated Learning
- URL: http://arxiv.org/abs/2109.02053v1
- Date: Sun, 5 Sep 2021 12:17:00 GMT
- Title: GTG-Shapley: Efficient and Accurate Participant Contribution Evaluation
in Federated Learning
- Authors: Zelei Liu, Yuanyuan Chen, Han Yu, Yang Liu and Lizhen Cui
- Abstract summary: Federated Learning (FL) bridges the gap between collaborative machine learning and preserving data privacy.
It is essential to fairly evaluate participants' contribution to the performance of the final FL model without exposing their private data.
We propose the Guided Truncation Gradient Shapley approach to address this challenge.
- Score: 25.44023017628766
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated Learning (FL) bridges the gap between collaborative machine
learning and preserving data privacy. To sustain the long-term operation of an
FL ecosystem, it is important to attract high quality data owners with
appropriate incentive schemes. As an important building block of such incentive
schemes, it is essential to fairly evaluate participants' contribution to the
performance of the final FL model without exposing their private data. Shapley
Value (SV)-based techniques have been widely adopted to provide fair evaluation
of FL participant contributions. However, existing approaches incur significant
computation costs, making them difficult to apply in practice. In this paper,
we propose the Guided Truncation Gradient Shapley (GTG-Shapley) approach to
address this challenge. It reconstructs FL models from gradient updates for SV
calculation instead of repeatedly training with different combinations of FL
participants. In addition, we design a guided Monte Carlo sampling approach
combined with within-round and between-round truncation to further reduce the
number of model reconstructions and evaluations required, through extensive
experiments under diverse realistic data distribution settings. The results
demonstrate that GTG-Shapley can closely approximate actual Shapley values,
while significantly increasing computational efficiency compared to the state
of the art, especially under non-i.i.d. settings.
Related papers
- Can We Theoretically Quantify the Impacts of Local Updates on the Generalization Performance of Federated Learning? [50.03434441234569]
Federated Learning (FL) has gained significant popularity due to its effectiveness in training machine learning models across diverse sites without requiring direct data sharing.
While various algorithms have shown that FL with local updates is a communication-efficient distributed learning framework, the generalization performance of FL with local updates has received comparatively less attention.
arXiv Detail & Related papers (2024-09-05T19:00:18Z) - Redefining Contributions: Shapley-Driven Federated Learning [3.9539878659683363]
Federated learning (FL) has emerged as a pivotal approach in machine learning.
It is challenging to ensure global model convergence when participants do not contribute equally and/or honestly.
This paper proposes a novel contribution assessment method called ShapFed for fine-grained evaluation of participant contributions in FL.
arXiv Detail & Related papers (2024-06-01T22:40:31Z) - Don't Forget What I did?: Assessing Client Contributions in Federated
Learning [9.56869689239781]
Federated Learning (FL) is a collaborative machine learning (ML) approach, where multiple clients participate in training an ML model without exposing the private data.
We propose a history-aware game-theoretic framework, called FLContrib, to assess client contributions when a subset of clients participate in each epoch of FL training.
To demonstrate the benefits of history-aware client contributions, we apply FLContrib to detect dishonest clients conducting data poisoning in FL training.
arXiv Detail & Related papers (2024-03-11T20:39:32Z) - Take History as a Mirror in Heterogeneous Federated Learning [9.187993085263209]
Federated Learning (FL) allows several clients to cooperatively train machine learning models without disclosing the raw data.
In this work, we propose a novel asynchronous FL framework called Federated Historical Learning (FedHist)
FedHist effectively addresses the challenges posed by both Non-IID data and gradient staleness.
arXiv Detail & Related papers (2023-12-16T11:40:49Z) - Adaptive Model Pruning and Personalization for Federated Learning over
Wireless Networks [72.59891661768177]
Federated learning (FL) enables distributed learning across edge devices while protecting data privacy.
We consider a FL framework with partial model pruning and personalization to overcome these challenges.
This framework splits the learning model into a global part with model pruning shared with all devices to learn data representations and a personalized part to be fine-tuned for a specific device.
arXiv Detail & Related papers (2023-09-04T21:10:45Z) - Integrating Local Real Data with Global Gradient Prototypes for
Classifier Re-Balancing in Federated Long-Tailed Learning [60.41501515192088]
Federated Learning (FL) has become a popular distributed learning paradigm that involves multiple clients training a global model collaboratively.
The data samples usually follow a long-tailed distribution in the real world, and FL on the decentralized and long-tailed data yields a poorly-behaved global model.
In this work, we integrate the local real data with the global gradient prototypes to form the local balanced datasets.
arXiv Detail & Related papers (2023-01-25T03:18:10Z) - Efficient Split-Mix Federated Learning for On-Demand and In-Situ
Customization [107.72786199113183]
Federated learning (FL) provides a distributed learning framework for multiple participants to collaborate learning without sharing raw data.
In this paper, we propose a novel Split-Mix FL strategy for heterogeneous participants that, once training is done, provides in-situ customization of model sizes and robustness.
arXiv Detail & Related papers (2022-03-18T04:58:34Z) - Federated Robustness Propagation: Sharing Adversarial Robustness in
Federated Learning [98.05061014090913]
Federated learning (FL) emerges as a popular distributed learning schema that learns from a set of participating users without requiring raw data to be shared.
adversarial training (AT) provides a sound solution for centralized learning, extending its usage for FL users has imposed significant challenges.
We show that existing FL techniques cannot effectively propagate adversarial robustness among non-iid users.
We propose a simple yet effective propagation approach that transfers robustness through carefully designed batch-normalization statistics.
arXiv Detail & Related papers (2021-06-18T15:52:33Z) - FedScale: Benchmarking Model and System Performance of Federated
Learning [4.1617240682257925]
FedScale is a set of challenging and realistic benchmark datasets for federated learning (FL) research.
FedScale is open-source with permissive licenses and actively maintained.
arXiv Detail & Related papers (2021-05-24T15:55:27Z) - A Principled Approach to Data Valuation for Federated Learning [73.19984041333599]
Federated learning (FL) is a popular technique to train machine learning (ML) models on decentralized data sources.
The Shapley value (SV) defines a unique payoff scheme that satisfies many desiderata for a data value notion.
This paper proposes a variant of the SV amenable to FL, which we call the federated Shapley value.
arXiv Detail & Related papers (2020-09-14T04:37:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.