FedV: Privacy-Preserving Federated Learning over Vertically Partitioned
Data
- URL: http://arxiv.org/abs/2103.03918v1
- Date: Fri, 5 Mar 2021 19:59:29 GMT
- Title: FedV: Privacy-Preserving Federated Learning over Vertically Partitioned
Data
- Authors: Runhua Xu, Nathalie Baracaldo, Yi Zhou, Ali Anwar, James Joshi, Heiko
Ludwig
- Abstract summary: Federated learning (FL) has been proposed to allow collaborative training of machine learning (ML) models among multiple parties.
We propose FedV, a framework for secure gradient computation in vertical settings for several widely used ML models.
We show a reduction of 10%-70% of training time and 80% to 90% in data transfer with respect to the state-of-the-art approaches.
- Score: 12.815996963583641
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Federated learning (FL) has been proposed to allow collaborative training of
machine learning (ML) models among multiple parties where each party can keep
its data private. In this paradigm, only model updates, such as model weights
or gradients, are shared. Many existing approaches have focused on horizontal
FL, where each party has the entire feature set and labels in the training data
set. However, many real scenarios follow a vertically-partitioned FL setup,
where a complete feature set is formed only when all the datasets from the
parties are combined, and the labels are only available to a single party.
Privacy-preserving vertical FL is challenging because complete sets of labels
and features are not owned by one entity. Existing approaches for vertical FL
require multiple peer-to-peer communications among parties, leading to lengthy
training times, and are restricted to (approximated) linear models and just two
parties. To close this gap, we propose FedV, a framework for secure gradient
computation in vertical settings for several widely used ML models such as
linear models, logistic regression, and support vector machines. FedV removes
the need for peer-to-peer communication among parties by using functional
encryption schemes; this allows FedV to achieve faster training times. It also
works for larger and changing sets of parties. We empirically demonstrate the
applicability for multiple types of ML models and show a reduction of 10%-70%
of training time and 80% to 90% in data transfer with respect to the
state-of-the-art approaches.
Related papers
- Federated Transformer: Multi-Party Vertical Federated Learning on Practical Fuzzily Linked Data [27.073959939557362]
We introduce the Federated Transformer (FeT), a novel framework that supports multi-party fuzzy VFL with fuzzy identifiers.
Our experiments demonstrate that the FeT surpasses the baseline models by up to 46% in terms of accuracy when scaled to 50 parties.
In two-party fuzzy VFL settings, FeT also shows improved performance and privacy over cutting-edge VFL models.
arXiv Detail & Related papers (2024-10-23T16:00:14Z) - MultiConfederated Learning: Inclusive Non-IID Data handling with Decentralized Federated Learning [1.2726316791083532]
Federated Learning (FL) has emerged as a prominent privacy-preserving technique for enabling use cases like confidential clinical machine learning.
FL operates by aggregating models trained by remote devices which owns the data.
We propose MultiConfederated Learning: a decentralized FL framework which is designed to handle non-IID data.
arXiv Detail & Related papers (2024-04-20T16:38:26Z) - Towards More Suitable Personalization in Federated Learning via
Decentralized Partial Model Training [67.67045085186797]
Almost all existing systems have to face large communication burdens if the central FL server fails.
It personalizes the "right" in the deep models by alternately updating the shared and personal parameters.
To further promote the shared parameters aggregation process, we propose DFed integrating the local Sharpness Miniization.
arXiv Detail & Related papers (2023-05-24T13:52:18Z) - Quadratic Functional Encryption for Secure Training in Vertical
Federated Learning [26.188083606166806]
Vertical federated learning (VFL) enables the collaborative training of machine learning (ML) models in settings where the data is distributed amongst multiple parties.
In VFL, the labels are available to a single party and the complete feature set is formed only when data from all parties is combined.
Recently, Xu et al. proposed a new framework called FedV for secure gradient computation for VFL using multi-input functional encryption.
arXiv Detail & Related papers (2023-05-15T05:31:35Z) - Scalable Collaborative Learning via Representation Sharing [53.047460465980144]
Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
arXiv Detail & Related papers (2022-11-20T10:49:22Z) - Federated Learning from Only Unlabeled Data with
Class-Conditional-Sharing Clients [98.22390453672499]
Supervised federated learning (FL) enables multiple clients to share the trained model without sharing their labeled data.
We propose federation of unsupervised learning (FedUL), where the unlabeled data are transformed into surrogate labeled data for each of the clients.
arXiv Detail & Related papers (2022-04-07T09:12:00Z) - A Vertical Federated Learning Framework for Horizontally Partitioned
Labels [12.433809611989155]
Most existing vertical federated learning methods have a strong assumption that at least one party holds the complete set of labels of all data samples.
Existing vertical federated learning methods can only utilize partial labels, which may lead to inadequate model update in end-to-end backpropagation.
We propose a novel vertical federated learning framework named Cascade Vertical Federated Learning (CVFL) to fully utilize all horizontally partitioned labels to train neural networks with privacy-preservation.
arXiv Detail & Related papers (2021-06-18T11:10:11Z) - FedH2L: Federated Learning with Model and Statistical Heterogeneity [75.61234545520611]
Federated learning (FL) enables distributed participants to collectively learn a strong global model without sacrificing their individual data privacy.
We introduce FedH2L, which is agnostic to both the model architecture and robust to different data distributions across participants.
In contrast to approaches sharing parameters or gradients, FedH2L relies on mutual distillation, exchanging only posteriors on a shared seed set between participants in a decentralized manner.
arXiv Detail & Related papers (2021-01-27T10:10:18Z) - FedCVT: Semi-supervised Vertical Federated Learning with Cross-view Training [9.638604434238882]
Federated Cross-view Training (FedCVT) is a semi-supervised learning approach that improves the performance of a vertical federated learning model.
FedCVT does not require parties to share their original data and model parameters, thus preserving data privacy.
arXiv Detail & Related papers (2020-08-25T06:20:31Z) - Ensemble Distillation for Robust Model Fusion in Federated Learning [72.61259487233214]
Federated Learning (FL) is a machine learning setting where many devices collaboratively train a machine learning model.
In most of the current training schemes the central model is refined by averaging the parameters of the server model and the updated parameters from the client side.
We propose ensemble distillation for model fusion, i.e. training the central classifier through unlabeled data on the outputs of the models from the clients.
arXiv Detail & Related papers (2020-06-12T14:49:47Z) - UVeQFed: Universal Vector Quantization for Federated Learning [179.06583469293386]
Federated learning (FL) is an emerging approach to train such learning models without requiring the users to share their possibly private labeled data.
In FL, each user trains its copy of the learning model locally. The server then collects the individual updates and aggregates them into a global model.
We show that combining universal vector quantization methods with FL yields a decentralized training system in which the compression of the trained models induces only a minimum distortion.
arXiv Detail & Related papers (2020-06-05T07:10:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.