Scalable Collaborative Learning via Representation Sharing
- URL: http://arxiv.org/abs/2211.10943v1
- Date: Sun, 20 Nov 2022 10:49:22 GMT
- Title: Scalable Collaborative Learning via Representation Sharing
- Authors: Fr\'ed\'eric Berdoz, Abhishek Singh, Martin Jaggi and Ramesh Raskar
- Abstract summary: Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
- Score: 53.047460465980144
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Privacy-preserving machine learning has become a key conundrum for
multi-party artificial intelligence. Federated learning (FL) and Split Learning
(SL) are two frameworks that enable collaborative learning while keeping the
data private (on device). In FL, each data holder trains a model locally and
releases it to a central server for aggregation. In SL, the clients must
release individual cut-layer activations (smashed data) to the server and wait
for its response (during both inference and back propagation). While relevant
in several settings, both of these schemes have a high communication cost, rely
on server-level computation algorithms and do not allow for tunable levels of
collaboration. In this work, we present a novel approach for privacy-preserving
machine learning, where the clients collaborate via online knowledge
distillation using a contrastive loss (contrastive w.r.t. the labels). The goal
is to ensure that the participants learn similar features on similar classes
without sharing their input data. To do so, each client releases averaged last
hidden layer activations of similar labels to a central server that only acts
as a relay (i.e., is not involved in the training or aggregation of the
models). Then, the clients download these last layer activations (feature
representations) of the ensemble of users and distill their knowledge in their
personal model using a contrastive objective. For cross-device applications
(i.e., small local datasets and limited computational capacity), this approach
increases the utility of the models compared to independent learning and other
federated knowledge distillation (FD) schemes, is communication efficient and
is scalable with the number of clients. We prove theoretically that our
framework is well-posed, and we benchmark its performance against standard FD
and FL on various datasets using different model architectures.
Related papers
- Personalized federated learning based on feature fusion [2.943623084019036]
Federated learning enables distributed clients to collaborate on training while storing their data locally to protect client privacy.
We propose a personalized federated learning approach called pFedPM.
In our process, we replace traditional gradient uploading with feature uploading, which helps reduce communication costs and allows for heterogeneous client models.
arXiv Detail & Related papers (2024-06-24T12:16:51Z) - Communication Efficient ConFederated Learning: An Event-Triggered SAGA
Approach [67.27031215756121]
Federated learning (FL) is a machine learning paradigm that targets model training without gathering the local data over various data sources.
Standard FL, which employs a single server, can only support a limited number of users, leading to degraded learning capability.
In this work, we consider a multi-server FL framework, referred to as emphConfederated Learning (CFL) in order to accommodate a larger number of users.
arXiv Detail & Related papers (2024-02-28T03:27:10Z) - PFSL: Personalized & Fair Split Learning with Data & Label Privacy for
thin clients [0.5144809478361603]
PFSL is a new framework of distributed split learning where a large number of thin clients perform transfer learning in parallel.
We implement a lightweight step of personalization of client models to provide high performance for their respective data distributions.
Our accuracy far exceeds that of current algorithms SL and is very close to that of centralized learning on several real-life benchmarks.
arXiv Detail & Related papers (2023-03-19T10:38:29Z) - Subspace based Federated Unlearning [75.90552823500633]
Federated unlearning (FL) aims to remove a specified target client's contribution in FL to satisfy the user's right to be forgotten.
Most existing federated unlearning algorithms require the server to store the history of the parameter updates.
We propose a simple-yet-effective subspace based federated unlearning method, dubbed SFU, that lets the global model perform gradient ascent.
arXiv Detail & Related papers (2023-02-24T04:29:44Z) - FedHP: Heterogeneous Federated Learning with Privacy-preserving [0.0]
Federated learning is a distributed machine learning environment, which ensures that clients complete collaborative training without sharing private data, only by exchanging parameters.
We propose a novel federated learning method, which consists of the pre-trained model as the backbone and fully connected layers as the head.
By sharing the embedding vector of classes, instead of parameters based on gradient space, clients can better adapt to private data, and it is more efficient in the communication between the server and clients.
arXiv Detail & Related papers (2023-01-27T13:32:17Z) - Distributed Unsupervised Visual Representation Learning with Fused
Features [13.935997509072669]
Federated learning (FL) enables distributed clients to learn a shared model for prediction while keeping the training data local on each client.
We propose a federated contrastive learning framework consisting of two approaches: feature fusion and neighborhood matching.
It outperforms other methods by 11% on IID data and matches the performance of centralized learning.
arXiv Detail & Related papers (2021-11-21T08:36:31Z) - An Expectation-Maximization Perspective on Federated Learning [75.67515842938299]
Federated learning describes the distributed training of models across multiple clients while keeping the data private on-device.
In this work, we view the server-orchestrated federated learning process as a hierarchical latent variable model where the server provides the parameters of a prior distribution over the client-specific model parameters.
We show that with simple Gaussian priors and a hard version of the well known Expectation-Maximization (EM) algorithm, learning in such a model corresponds to FedAvg, the most popular algorithm for the federated learning setting.
arXiv Detail & Related papers (2021-11-19T12:58:59Z) - Multimodal Federated Learning [9.081857621783811]
In many applications, such as smart homes with IoT devices, local data on clients are generated from different modalities.
Existing federated learning systems only work on local data from a single modality, which limits the scalability of the systems.
We propose a multimodal and semi-supervised federated learning framework that trains autoencoders to extract shared or correlated representations from different local data modalities on clients.
arXiv Detail & Related papers (2021-09-10T12:32:46Z) - FedKD: Communication Efficient Federated Learning via Knowledge
Distillation [56.886414139084216]
Federated learning is widely used to learn intelligent models from decentralized data.
In federated learning, clients need to communicate their local model updates in each iteration of model learning.
We propose a communication efficient federated learning method based on knowledge distillation.
arXiv Detail & Related papers (2021-08-30T15:39:54Z) - Ensemble Distillation for Robust Model Fusion in Federated Learning [72.61259487233214]
Federated Learning (FL) is a machine learning setting where many devices collaboratively train a machine learning model.
In most of the current training schemes the central model is refined by averaging the parameters of the server model and the updated parameters from the client side.
We propose ensemble distillation for model fusion, i.e. training the central classifier through unlabeled data on the outputs of the models from the clients.
arXiv Detail & Related papers (2020-06-12T14:49:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.