SemiFed: Semi-supervised Federated Learning with Consistency and
Pseudo-Labeling
- URL: http://arxiv.org/abs/2108.09412v1
- Date: Sat, 21 Aug 2021 01:14:27 GMT
- Title: SemiFed: Semi-supervised Federated Learning with Consistency and
Pseudo-Labeling
- Authors: Haowen Lin, Jian Lou, Li Xiong, Cyrus Shahabi
- Abstract summary: Federated learning enables multiple clients, such as mobile phones and organizations, to collaboratively learn a shared model for prediction.
In this work, we focus on a new scenario for cross-silo federated learning, where data samples of each client are partially labeled.
We propose a new framework dubbed SemiFed that unifies two dominant approaches for semi-supervised learning: consistency regularization and pseudo-labeling.
- Score: 14.737638416823772
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning enables multiple clients, such as mobile phones and
organizations, to collaboratively learn a shared model for prediction while
protecting local data privacy. However, most recent research and applications
of federated learning assume that all clients have fully labeled data, which is
impractical in real-world settings. In this work, we focus on a new scenario
for cross-silo federated learning, where data samples of each client are
partially labeled. We borrow ideas from semi-supervised learning methods where
a large amount of unlabeled data is utilized to improve the model's accuracy
despite limited access to labeled examples. We propose a new framework dubbed
SemiFed that unifies two dominant approaches for semi-supervised learning:
consistency regularization and pseudo-labeling. SemiFed first applies advanced
data augmentation techniques to enforce consistency regularization and then
generates pseudo-labels using the model's predictions during training. SemiFed
takes advantage of the federation so that for a given image, the pseudo-label
holds only if multiple models from different clients produce a high-confidence
prediction and agree on the same label. Extensive experiments on two image
benchmarks demonstrate the effectiveness of our approach under both homogeneous
and heterogeneous data distribution settings
Related papers
- Memory Consistency Guided Divide-and-Conquer Learning for Generalized
Category Discovery [56.172872410834664]
Generalized category discovery (GCD) aims at addressing a more realistic and challenging setting of semi-supervised learning.
We propose a Memory Consistency guided Divide-and-conquer Learning framework (MCDL)
Our method outperforms state-of-the-art models by a large margin on both seen and unseen classes of the generic image recognition.
arXiv Detail & Related papers (2024-01-24T09:39:45Z) - Rethinking Semi-Supervised Federated Learning: How to co-train
fully-labeled and fully-unlabeled client imaging data [6.322831694506287]
Isolated Federated Learning (IsoFed) is a learning scheme specifically designed for semi-supervised federated learning (SSFL)
We propose a novel learning scheme specifically designed for SSFL that circumvents the problem by avoiding simple averaging of supervised and semi-supervised models together.
In particular, our training approach consists of two parts - (a) isolated aggregation of labeled and unlabeled client models, and (b) local self-supervised pretraining of isolated global models in all clients.
arXiv Detail & Related papers (2023-10-28T20:41:41Z) - JointMatch: A Unified Approach for Diverse and Collaborative
Pseudo-Labeling to Semi-Supervised Text Classification [65.268245109828]
Semi-supervised text classification (SSTC) has gained increasing attention due to its ability to leverage unlabeled data.
Existing approaches based on pseudo-labeling suffer from the issues of pseudo-label bias and error accumulation.
We propose JointMatch, a holistic approach for SSTC that addresses these challenges by unifying ideas from recent semi-supervised learning.
arXiv Detail & Related papers (2023-10-23T05:43:35Z) - Federated Semi-Supervised Learning with Annotation Heterogeneity [57.12560313403097]
We propose a novel framework called Heterogeneously Annotated Semi-Supervised LEarning (HASSLE)
It is a dual-model framework with two models trained separately on labeled and unlabeled data.
The dual models can implicitly learn from both types of data across different clients, although each dual model is only trained locally on a single type of data.
arXiv Detail & Related papers (2023-03-04T16:04:49Z) - FedIL: Federated Incremental Learning from Decentralized Unlabeled Data
with Convergence Analysis [23.70951896315126]
This work considers the server with a small labeled dataset and intends to use unlabeled data in multiple clients for semi-supervised learning.
We propose a new framework with a generalized model, Federated Incremental Learning (FedIL), to address the problem of how to utilize labeled data in the server and unlabeled data in clients separately.
arXiv Detail & Related papers (2023-02-23T07:12:12Z) - When to Trust Aggregated Gradients: Addressing Negative Client Sampling
in Federated Learning [41.51682329500003]
We propose a novel learning rate adaptation mechanism to adjust the server learning rate for the aggregated gradient in each round.
We make theoretical deductions to find a meaningful and robust indicator that is positively related to the optimal server learning rate.
arXiv Detail & Related papers (2023-01-25T03:52:45Z) - Towards Semi-Supervised Deep Facial Expression Recognition with An
Adaptive Confidence Margin [92.76372026435858]
We learn an Adaptive Confidence Margin (Ada-CM) to fully leverage all unlabeled data for semi-supervised deep facial expression recognition.
All unlabeled samples are partitioned into two subsets by comparing their confidence scores with the adaptively learned confidence margin.
Our method achieves state-of-the-art performance, especially surpassing fully-supervised baselines in a semi-supervised manner.
arXiv Detail & Related papers (2022-03-23T11:43:29Z) - Federated Unsupervised Representation Learning [56.715917111878106]
We formulate a new problem in federated learning called Federated Unsupervised Representation Learning (FURL) to learn a common representation model without supervision.
FedCA is composed of two key modules: dictionary module to aggregate the representations of samples from each client and share with all clients for consistency of representation space and alignment module to align the representation of each client on a base model trained on a public data.
arXiv Detail & Related papers (2020-10-18T13:28:30Z) - Federated Semi-Supervised Learning with Inter-Client Consistency &
Disjoint Learning [78.88007892742438]
We study two essential scenarios of Federated Semi-Supervised Learning (FSSL) based on the location of the labeled data.
We propose a novel method to tackle the problems, which we refer to as Federated Matching (FedMatch)
arXiv Detail & Related papers (2020-06-22T09:43:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.