DBFed: Debiasing Federated Learning Framework based on
Domain-Independent
- URL: http://arxiv.org/abs/2307.05582v1
- Date: Mon, 10 Jul 2023 14:39:57 GMT
- Title: DBFed: Debiasing Federated Learning Framework based on
Domain-Independent
- Authors: Jiale Li, Zhixin Li, Yibo Wang, Yao Li, Lei Wang
- Abstract summary: We propose a debiasing federated learning framework based on domain-independent, which mitigates model bias by explicitly encoding sensitive attributes during client-side training.
This paper conducts experiments on three real datasets and uses five evaluation metrics of accuracy and fairness to quantify the effect of the model.
- Score: 15.639705798326213
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As digital transformation continues, enterprises are generating, managing,
and storing vast amounts of data, while artificial intelligence technology is
rapidly advancing. However, it brings challenges in information security and
data security. Data security refers to the protection of digital information
from unauthorized access, damage, theft, etc. throughout its entire life cycle.
With the promulgation and implementation of data security laws and the emphasis
on data security and data privacy by organizations and users,
Privacy-preserving technology represented by federated learning has a wide
range of application scenarios. Federated learning is a distributed machine
learning computing framework that allows multiple subjects to train joint
models without sharing data to protect data privacy and solve the problem of
data islands. However, the data among multiple subjects are independent of each
other, and the data differences in quality may cause fairness issues in
federated learning modeling, such as data bias among multiple subjects,
resulting in biased and discriminatory models. Therefore, we propose DBFed, a
debiasing federated learning framework based on domain-independent, which
mitigates model bias by explicitly encoding sensitive attributes during
client-side training. This paper conducts experiments on three real datasets
and uses five evaluation metrics of accuracy and fairness to quantify the
effect of the model. Most metrics of DBFed exceed those of the other three
comparative methods, fully demonstrating the debiasing effect of DBFed.
Related papers
- FewFedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning [54.26614091429253]
Federated instruction tuning (FedIT) is a promising solution, by consolidating collaborative training across multiple data owners.
FedIT encounters limitations such as scarcity of instructional data and risk of exposure to training data extraction attacks.
We propose FewFedPIT, designed to simultaneously enhance privacy protection and model performance of federated few-shot learning.
arXiv Detail & Related papers (2024-03-10T08:41:22Z) - Privacy-Enhancing Collaborative Information Sharing through Federated
Learning -- A Case of the Insurance Industry [1.8092553911119764]
The report demonstrates the benefits of harnessing the value of Federated Learning (FL) to learn a single model across multiple insurance industry datasets.
FL addresses two of the most pressing concerns: limited data volume and data variety, which are caused by privacy concerns.
During each round of FL, collaborators compute improvements on the model using their local private data, and these insights are combined to update a global model.
arXiv Detail & Related papers (2024-02-22T21:46:24Z) - Federated Learning Empowered by Generative Content [55.576885852501775]
Federated learning (FL) enables leveraging distributed private data for model training in a privacy-preserving way.
We propose a novel FL framework termed FedGC, designed to mitigate data heterogeneity issues by diversifying private data with generative content.
We conduct a systematic empirical study on FedGC, covering diverse baselines, datasets, scenarios, and modalities.
arXiv Detail & Related papers (2023-12-10T07:38:56Z) - Approximate, Adapt, Anonymize (3A): a Framework for Privacy Preserving
Training Data Release for Machine Learning [3.29354893777827]
We introduce a data release framework, 3A (Approximate, Adapt, Anonymize), to maximize data utility for machine learning.
We present experimental evidence showing minimal discrepancy between performance metrics of models trained on real versus privatized datasets.
arXiv Detail & Related papers (2023-07-04T18:37:11Z) - Benchmarking FedAvg and FedCurv for Image Classification Tasks [1.376408511310322]
This paper focuses on the problem of statistical heterogeneity of the data in the same federated network.
Several Federated Learning algorithms, such as FedAvg, FedProx and Federated Curvature (FedCurv) have already been proposed.
As a side product of this work, we release the non-IID version of the datasets we used so to facilitate further comparisons from the FL community.
arXiv Detail & Related papers (2023-03-31T10:13:01Z) - Adversarial Representation Sharing: A Quantitative and Secure
Collaborative Learning Framework [3.759936323189418]
We find representation learning has unique advantages in collaborative learning due to the lower communication overhead and task-independency.
We present ARS, a collaborative learning framework wherein users share representations of data to train models.
We demonstrate that our mechanism is effective against model inversion attacks, and achieves a balance between privacy and utility.
arXiv Detail & Related papers (2022-03-27T13:29:15Z) - Representative & Fair Synthetic Data [68.8204255655161]
We present a framework to incorporate fairness constraints into the self-supervised learning process.
We generate a representative as well as fair version of the UCI Adult census data set.
We consider representative & fair synthetic data a promising future building block to teach algorithms not on historic worlds, but rather on the worlds that we strive to live in.
arXiv Detail & Related papers (2021-04-07T09:19:46Z) - Decentralized Federated Learning Preserves Model and Data Privacy [77.454688257702]
We propose a fully decentralized approach, which allows to share knowledge between trained models.
Students are trained on the output of their teachers via synthetically generated input data.
The results show that an untrained student model, trained on the teachers output reaches comparable F1-scores as the teacher.
arXiv Detail & Related papers (2021-02-01T14:38:54Z) - Vertical federated learning based on DFP and BFGS [0.0]
We propose a novel vertical federated learning framework based on the DFP and the BFGS(denoted as BDFL)
We perform experiments using real datasets to test efficiency of BDFL framework.
arXiv Detail & Related papers (2021-01-23T06:15:04Z) - Trustworthy AI [75.99046162669997]
Brittleness to minor adversarial changes in the input data, ability to explain the decisions, address the bias in their training data, are some of the most prominent limitations.
We propose the tutorial on Trustworthy AI to address six critical issues in enhancing user and public trust in AI systems.
arXiv Detail & Related papers (2020-11-02T20:04:18Z) - WAFFLe: Weight Anonymized Factorization for Federated Learning [88.44939168851721]
In domains where data are sensitive or private, there is great value in methods that can learn in a distributed manner without the data ever leaving the local devices.
We propose Weight Anonymized Factorization for Federated Learning (WAFFLe), an approach that combines the Indian Buffet Process with a shared dictionary of weight factors for neural networks.
arXiv Detail & Related papers (2020-08-13T04:26:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.