PyVertical: A Vertical Federated Learning Framework for Multi-headed
SplitNN
- URL: http://arxiv.org/abs/2104.00489v1
- Date: Thu, 1 Apr 2021 14:21:33 GMT
- Title: PyVertical: A Vertical Federated Learning Framework for Multi-headed
SplitNN
- Authors: Daniele Romanini, Adam James Hall, Pavlos Papadopoulos, Tom Titcombe,
Abbas Ismail, Tudor Cebere, Robert Sandmann, Robin Roehm, Michael A. Hoeh
- Abstract summary: PyVertical is a framework supporting vertical federated learning using split neural networks.
We present the training of a simple dual-headed split neural network for a MNIST classification task, with data samples vertically distributed across two data owners and a data scientist.
- Score: 4.552318473944852
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce PyVertical, a framework supporting vertical federated learning
using split neural networks. The proposed framework allows a data scientist to
train neural networks on data features vertically partitioned across multiple
owners while keeping raw data on an owner's device. To link entities shared
across different datasets' partitions, we use Private Set Intersection on IDs
associated with data points. To demonstrate the validity of the proposed
framework, we present the training of a simple dual-headed split neural network
for a MNIST classification task, with data samples vertically distributed
across two data owners and a data scientist.
Related papers
- Neural Network Training on Encrypted Data with TFHE [0.0]
We present an approach to outsourcing of training neural networks while preserving data confidentiality from malicious parties.
We use fully homomorphic encryption to build a unified training approach that works on encrypted data and learns quantized neural network models.
arXiv Detail & Related papers (2024-01-29T13:07:08Z) - Scalable Collaborative Learning via Representation Sharing [53.047460465980144]
Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
arXiv Detail & Related papers (2022-11-20T10:49:22Z) - Vertical Federated Principal Component Analysis and Its Kernel Extension
on Feature-wise Distributed Data [35.72930187906397]
This paper will study the unsupervised federated learning under the vertically partitioned dataset setting.
We propose the federated principal component analysis for vertically partitioned dataset (VFedPCA) method.
We also take advantage of the nonlinear dimensionality reduction and propose the vertical federated advanced kernel principal component analysis (VFedAKPCA) method.
arXiv Detail & Related papers (2022-03-03T14:58:29Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Cross-Silo Federated Learning for Multi-Tier Networks with Vertical and Horizontal Data Partitioning [20.410494162333972]
Tiered Decentralized Coordinate Descent (TDCD) is a decentralized training algorithm for two-tiered communication networks.
We present a theoretical analysis of our algorithm and show the dependence of the convergence rate on the number of vertical partitions and the number of local updates.
arXiv Detail & Related papers (2021-08-19T22:01:04Z) - Multi-VFL: A Vertical Federated Learning System for Multiple Data and
Label Owners [10.507522234243021]
We propose a novel method, Multi Vertical Federated Learning (Multi-VFL), to train VFL models when there are multiple data and label owners.
Our results show that using adaptive datasets for model aggregation fastens convergence and improves accuracy.
arXiv Detail & Related papers (2021-06-10T03:00:57Z) - Multi-Tier Federated Learning for Vertically Partitioned Data [2.817412580574242]
We consider decentralized model training in tiered datasets networks.
Our network model consists of a set of silos, each holding a vertical partition of the data.
arXiv Detail & Related papers (2021-02-06T17:34:14Z) - Graph-Based Neural Network Models with Multiple Self-Supervised
Auxiliary Tasks [79.28094304325116]
Graph Convolutional Networks are among the most promising approaches for capturing relationships among structured data points.
We propose three novel self-supervised auxiliary tasks to train graph-based neural network models in a multi-task fashion.
arXiv Detail & Related papers (2020-11-14T11:09:51Z) - Federated Doubly Stochastic Kernel Learning for Vertically Partitioned
Data [93.76907759950608]
We propose a doubly kernel learning algorithm for vertically partitioned data.
We show that FDSKL is significantly faster than state-of-the-art federated learning methods when dealing with kernels.
arXiv Detail & Related papers (2020-08-14T05:46:56Z) - Pre-Trained Models for Heterogeneous Information Networks [57.78194356302626]
We propose a self-supervised pre-training and fine-tuning framework, PF-HIN, to capture the features of a heterogeneous information network.
PF-HIN consistently and significantly outperforms state-of-the-art alternatives on each of these tasks, on four datasets.
arXiv Detail & Related papers (2020-07-07T03:36:28Z) - Neural Data Server: A Large-Scale Search Engine for Transfer Learning
Data [78.74367441804183]
We introduce Neural Data Server (NDS), a large-scale search engine for finding the most useful transfer learning data to the target domain.
NDS consists of a dataserver which indexes several large popular image datasets, and aims to recommend data to a client.
We show the effectiveness of NDS in various transfer learning scenarios, demonstrating state-of-the-art performance on several target datasets.
arXiv Detail & Related papers (2020-01-09T01:21:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.