Privacy-Preserving Self-Taught Federated Learning for Heterogeneous Data
- URL: http://arxiv.org/abs/2102.05883v1
- Date: Thu, 11 Feb 2021 08:07:51 GMT
- Title: Privacy-Preserving Self-Taught Federated Learning for Heterogeneous Data
- Authors: Kai-Fung Chu, Lintao Zhang
- Abstract summary: Federated learning (FL) was proposed to enable joint training of a deep learning model using the local data in each party without revealing the data to others.
In this work, we propose an FL method called self-taught federated learning to address the aforementioned issues.
In this method, only latent variables are transmitted to other parties for model training, while privacy is preserved by storing the data and parameters of activations, weights, and biases locally.
- Score: 6.545317180430584
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Many application scenarios call for training a machine learning model among
multiple participants. Federated learning (FL) was proposed to enable joint
training of a deep learning model using the local data in each party without
revealing the data to others. Among various types of FL methods, vertical FL is
a category to handle data sources with the same ID space and different feature
spaces. However, existing vertical FL methods suffer from limitations such as
restrictive neural network structure, slow training speed, and often lack the
ability to take advantage of data with unmatched IDs. In this work, we propose
an FL method called self-taught federated learning to address the
aforementioned issues, which uses unsupervised feature extraction techniques
for distributed supervised deep learning tasks. In this method, only latent
variables are transmitted to other parties for model training, while privacy is
preserved by storing the data and parameters of activations, weights, and
biases locally. Extensive experiments are performed to evaluate and demonstrate
the validity and efficiency of the proposed method.
Related papers
- A review on different techniques used to combat the non-IID and
heterogeneous nature of data in FL [0.0]
Federated Learning (FL) is a machine-learning approach enabling collaborative model training across multiple edge devices.
The significance of FL is particularly pronounced in industries such as healthcare and finance, where data privacy holds paramount importance.
This report delves into the issues arising from non-IID and heterogeneous data and explores current algorithms designed to address these challenges.
arXiv Detail & Related papers (2024-01-01T16:34:00Z) - FedSym: Unleashing the Power of Entropy for Benchmarking the Algorithms
for Federated Learning [1.4656078321003647]
Federated learning (FL) is a decentralized machine learning approach where independent learners process data privately.
We study the currently popular data partitioning techniques and visualize their main disadvantages.
We propose a method that leverages entropy and symmetry to construct 'the most challenging' and controllable data distributions.
arXiv Detail & Related papers (2023-10-11T18:39:08Z) - Federated Learning with Privacy-Preserving Ensemble Attention
Distillation [63.39442596910485]
Federated Learning (FL) is a machine learning paradigm where many local nodes collaboratively train a central model while keeping the training data decentralized.
We propose a privacy-preserving FL framework leveraging unlabeled public data for one-way offline knowledge distillation.
Our technique uses decentralized and heterogeneous local data like existing FL approaches, but more importantly, it significantly reduces the risk of privacy leakage.
arXiv Detail & Related papers (2022-10-16T06:44:46Z) - FEDIC: Federated Learning on Non-IID and Long-Tailed Data via Calibrated
Distillation [54.2658887073461]
Dealing with non-IID data is one of the most challenging problems for federated learning.
This paper studies the joint problem of non-IID and long-tailed data in federated learning and proposes a corresponding solution called Federated Ensemble Distillation with Imbalance (FEDIC)
FEDIC uses model ensemble to take advantage of the diversity of models trained on non-IID data.
arXiv Detail & Related papers (2022-04-30T06:17:36Z) - Efficient Split-Mix Federated Learning for On-Demand and In-Situ
Customization [107.72786199113183]
Federated learning (FL) provides a distributed learning framework for multiple participants to collaborate learning without sharing raw data.
In this paper, we propose a novel Split-Mix FL strategy for heterogeneous participants that, once training is done, provides in-situ customization of model sizes and robustness.
arXiv Detail & Related papers (2022-03-18T04:58:34Z) - Local Learning Matters: Rethinking Data Heterogeneity in Federated
Learning [61.488646649045215]
Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices)
arXiv Detail & Related papers (2021-11-28T19:03:39Z) - Communication-Efficient Hierarchical Federated Learning for IoT
Heterogeneous Systems with Imbalanced Data [42.26599494940002]
Federated learning (FL) is a distributed learning methodology that allows multiple nodes to cooperatively train a deep learning model.
This paper studies the potential of hierarchical FL in IoT heterogeneous systems.
It proposes an optimized solution for user assignment and resource allocation on multiple edge nodes.
arXiv Detail & Related papers (2021-07-14T08:32:39Z) - Personalized Federated Learning over non-IID Data for Indoor
Localization [17.03722514121803]
We consider the use of recent Federated Learning schemes to train a set of personalized models.
In this paper, we consider the use of recent FL schemes to train a set of personalized models that are then optimally fused through Bayesian rules.
arXiv Detail & Related papers (2021-07-09T03:31:16Z) - Federated Robustness Propagation: Sharing Adversarial Robustness in
Federated Learning [98.05061014090913]
Federated learning (FL) emerges as a popular distributed learning schema that learns from a set of participating users without requiring raw data to be shared.
adversarial training (AT) provides a sound solution for centralized learning, extending its usage for FL users has imposed significant challenges.
We show that existing FL techniques cannot effectively propagate adversarial robustness among non-iid users.
We propose a simple yet effective propagation approach that transfers robustness through carefully designed batch-normalization statistics.
arXiv Detail & Related papers (2021-06-18T15:52:33Z) - Mitigating Bias in Federated Learning [9.295028968787351]
In this paper, we discuss causes of bias in federated learning (FL)
We propose three pre-processing and in-processing methods to mitigate bias, without compromising data privacy.
We conduct experiments over several data distributions to analyze their effects on model performance, fairness metrics, and bias learning patterns.
arXiv Detail & Related papers (2020-12-04T08:04:12Z) - WAFFLe: Weight Anonymized Factorization for Federated Learning [88.44939168851721]
In domains where data are sensitive or private, there is great value in methods that can learn in a distributed manner without the data ever leaving the local devices.
We propose Weight Anonymized Factorization for Federated Learning (WAFFLe), an approach that combines the Indian Buffet Process with a shared dictionary of weight factors for neural networks.
arXiv Detail & Related papers (2020-08-13T04:26:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.