Collaborative Unsupervised Visual Representation Learning from
Decentralized Data
- URL: http://arxiv.org/abs/2108.06492v1
- Date: Sat, 14 Aug 2021 08:34:11 GMT
- Title: Collaborative Unsupervised Visual Representation Learning from
Decentralized Data
- Authors: Weiming Zhuang, Xin Gan, Yonggang Wen, Shuai Zhang, Shuai Yi
- Abstract summary: We propose a novel federated unsupervised learning framework, FedU.
In this framework, each party trains models from unlabeled data independently using contrastive learning with an online network and a target network.
FedU preserves data privacy as each party only has access to its raw data.
- Score: 34.06624704343615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised representation learning has achieved outstanding performances
using centralized data available on the Internet. However, the increasing
awareness of privacy protection limits sharing of decentralized unlabeled image
data that grows explosively in multiple parties (e.g., mobile phones and
cameras). As such, a natural problem is how to leverage these data to learn
visual representations for downstream tasks while preserving data privacy. To
address this problem, we propose a novel federated unsupervised learning
framework, FedU. In this framework, each party trains models from unlabeled
data independently using contrastive learning with an online network and a
target network. Then, a central server aggregates trained models and updates
clients' models with the aggregated model. It preserves data privacy as each
party only has access to its raw data. Decentralized data among multiple
parties are normally non-independent and identically distributed (non-IID),
leading to performance degradation. To tackle this challenge, we propose two
simple but effective methods: 1) We design the communication protocol to upload
only the encoders of online networks for server aggregation and update them
with the aggregated encoder; 2) We introduce a new module to dynamically decide
how to update predictors based on the divergence caused by non-IID. The
predictor is the other component of the online network. Extensive experiments
and ablations demonstrate the effectiveness and significance of FedU. It
outperforms training with only one party by over 5% and other methods by over
14% in linear and semi-supervised evaluation on non-IID data.
Related papers
- Federated Face Forgery Detection Learning with Personalized Representation [63.90408023506508]
Deep generator technology can produce high-quality fake videos that are indistinguishable, posing a serious social threat.
Traditional forgery detection methods directly centralized training on data.
The paper proposes a novel federated face forgery detection learning with personalized representation.
arXiv Detail & Related papers (2024-06-17T02:20:30Z) - Practical, Private Assurance of the Value of Collaboration via Fully Homomorphic Encryption [3.929854470352013]
Two parties wish to collaborate on their datasets.
One party is promised an improvement on its prediction model by incorporating data from the other party.
The parties would only wish to collaborate further if the updated model shows an improvement in accuracy.
arXiv Detail & Related papers (2023-10-04T03:47:21Z) - Benchmarking FedAvg and FedCurv for Image Classification Tasks [1.376408511310322]
This paper focuses on the problem of statistical heterogeneity of the data in the same federated network.
Several Federated Learning algorithms, such as FedAvg, FedProx and Federated Curvature (FedCurv) have already been proposed.
As a side product of this work, we release the non-IID version of the datasets we used so to facilitate further comparisons from the FL community.
arXiv Detail & Related papers (2023-03-31T10:13:01Z) - Scalable Collaborative Learning via Representation Sharing [53.047460465980144]
Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
arXiv Detail & Related papers (2022-11-20T10:49:22Z) - Rethinking Data Heterogeneity in Federated Learning: Introducing a New
Notion and Standard Benchmarks [65.34113135080105]
We show that not only the issue of data heterogeneity in current setups is not necessarily a problem but also in fact it can be beneficial for the FL participants.
Our observations are intuitive.
Our code is available at https://github.com/MMorafah/FL-SC-NIID.
arXiv Detail & Related papers (2022-09-30T17:15:19Z) - RelaySum for Decentralized Deep Learning on Heterogeneous Data [71.36228931225362]
In decentralized machine learning, workers compute model updates on their local data.
Because the workers only communicate with few neighbors without central coordination, these updates propagate progressively over the network.
This paradigm enables distributed training on networks without all-to-all connectivity, helping to protect data privacy as well as to reduce the communication cost of distributed training in data centers.
arXiv Detail & Related papers (2021-10-08T14:55:32Z) - Aggregation Delayed Federated Learning [20.973999078271483]
Federated learning is a distributed machine learning paradigm where multiple data owners (clients) collaboratively train one machine learning model while keeping data on their own devices.
Studies have found performance reduction with standard federated algorithms, such as FedAvg, on non-IID data.
Many existing works on handling non-IID data adopt the same aggregation framework as FedAvg and focus on improving model updates either on the server side or on clients.
In this work, we tackle this challenge by introducing redistribution rounds that delay the aggregation. We perform experiments on multiple tasks and show that the proposed framework significantly improves the performance on non-IID
arXiv Detail & Related papers (2021-08-17T04:06:10Z) - Joint Optimization in Edge-Cloud Continuum for Federated Unsupervised
Person Re-identification [24.305773593017932]
FedUReID is a federated unsupervised person ReID system to learn person ReID models without any labels while preserving privacy.
To tackle the problem that edges vary in data volumes and distributions, we personalize training in edges with joint optimization of cloud and edge.
Experiments on eight person ReID datasets demonstrate that FedUReID achieves higher accuracy but also reduces computation cost by 29%.
arXiv Detail & Related papers (2021-08-14T08:35:55Z) - Decentralized federated learning of deep neural networks on non-iid data [0.6335848702857039]
We tackle the non-problem of learning a personalized deep learning model in a decentralized setting.
We propose a method named Performance-Based Neighbor Selection (PENS) where clients with similar data detect each other and cooperate.
PENS is able to achieve higher accuracies as compared to strong baselines.
arXiv Detail & Related papers (2021-07-18T19:05:44Z) - Decentralized Federated Learning Preserves Model and Data Privacy [77.454688257702]
We propose a fully decentralized approach, which allows to share knowledge between trained models.
Students are trained on the output of their teachers via synthetically generated input data.
The results show that an untrained student model, trained on the teachers output reaches comparable F1-scores as the teacher.
arXiv Detail & Related papers (2021-02-01T14:38:54Z) - Decentralised Learning from Independent Multi-Domain Labels for Person
Re-Identification [69.29602103582782]
Deep learning has been successful for many computer vision tasks due to the availability of shared and centralised large-scale training data.
However, increasing awareness of privacy concerns poses new challenges to deep learning, especially for person re-identification (Re-ID)
We propose a novel paradigm called Federated Person Re-Identification (FedReID) to construct a generalisable global model (a central server) by simultaneously learning with multiple privacy-preserved local models (local clients)
This client-server collaborative learning process is iteratively performed under privacy control, enabling FedReID to realise decentralised learning without sharing distributed data nor collecting any
arXiv Detail & Related papers (2020-06-07T13:32:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.