Fair and efficient contribution valuation for vertical federated
learning
- URL: http://arxiv.org/abs/2201.02658v1
- Date: Fri, 7 Jan 2022 19:57:15 GMT
- Title: Fair and efficient contribution valuation for vertical federated
learning
- Authors: Zhenan Fan, Huang Fang, Zirui Zhou, Jian Pei, Michael P. Friedlander,
Yong Zhang
- Abstract summary: Federated learning is a popular technology for training machine learning models on distributed data sources without sharing data.
The Shapley value (SV) is a provably fair contribution valuation metric originated from cooperative game theory.
We propose a contribution valuation metric called vertical federated Shapley value (VerFedSV) based on SV.
- Score: 49.50442779626123
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning is a popular technology for training machine learning
models on distributed data sources without sharing data. Vertical federated
learning or feature-based federated learning applies to the cases that
different data sources share the same sample ID space but differ in feature
space. To ensure the data owners' long-term engagement, it is critical to
objectively assess the contribution from each data source and recompense them
accordingly. The Shapley value (SV) is a provably fair contribution valuation
metric originated from cooperative game theory. However, computing the SV
requires extensively retraining the model on each subset of data sources, which
causes prohibitively high communication costs in federated learning. We propose
a contribution valuation metric called vertical federated Shapley value
(VerFedSV) based on SV. We show that VerFedSV not only satisfies many desirable
properties for fairness but is also efficient to compute, and can be adapted to
both synchronous and asynchronous vertical federated learning algorithms. Both
theoretical analysis and extensive experimental results verify the fairness,
efficiency, and adaptability of VerFedSV.
Related papers
- Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Accelerating Fair Federated Learning: Adaptive Federated Adam [0.0]
When datasets are not independent and identically distributed (non-IID), models trained by naive federated algorithms may be biased towards certain participants.
This is known as the fairness problem in federated learning.
We propose Adaptive Federated Adam (AdaFedAdam) to accelerate fair federated learning with alleviated bias.
arXiv Detail & Related papers (2023-01-23T10:56:12Z) - FEDIC: Federated Learning on Non-IID and Long-Tailed Data via Calibrated
Distillation [54.2658887073461]
Dealing with non-IID data is one of the most challenging problems for federated learning.
This paper studies the joint problem of non-IID and long-tailed data in federated learning and proposes a corresponding solution called Federated Ensemble Distillation with Imbalance (FEDIC)
FEDIC uses model ensemble to take advantage of the diversity of models trained on non-IID data.
arXiv Detail & Related papers (2022-04-30T06:17:36Z) - Improving Fairness for Data Valuation in Federated Learning [39.61504568047234]
We propose a new measure called completed federated Shapley value to improve the fairness of federated Shapley value.
It is shown under mild conditions that this matrix is approximately low-rank by leveraging concepts and tools from optimization.
arXiv Detail & Related papers (2021-09-19T02:39:59Z) - Fairness-aware Agnostic Federated Learning [47.26747955026486]
We develop a fairness-aware agnostic federated learning framework (AgnosticFair) to deal with the challenge of unknown testing distribution.
We use kernel reweighing functions to assign a reweighing value on each training sample in both loss function and fairness constraint.
Built model can be directly applied to local sites as it guarantees fairness on local data distributions.
arXiv Detail & Related papers (2020-10-10T17:58:20Z) - A Principled Approach to Data Valuation for Federated Learning [73.19984041333599]
Federated learning (FL) is a popular technique to train machine learning (ML) models on decentralized data sources.
The Shapley value (SV) defines a unique payoff scheme that satisfies many desiderata for a data value notion.
This paper proposes a variant of the SV amenable to FL, which we call the federated Shapley value.
arXiv Detail & Related papers (2020-09-14T04:37:54Z) - Privacy-Preserving Asynchronous Federated Learning Algorithms for
Multi-Party Vertically Collaborative Learning [151.47900584193025]
We propose an asynchronous federated SGD (AFSGD-VP) algorithm and its SVRG and SAGA variants on the vertically partitioned data.
To the best of our knowledge, AFSGD-VP and its SVRG and SAGA variants are the first asynchronous federated learning algorithms for vertically partitioned data.
arXiv Detail & Related papers (2020-08-14T08:08:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.