Federated Semi-Supervised Learning with Inter-Client Consistency &
Disjoint Learning
- URL: http://arxiv.org/abs/2006.12097v3
- Date: Mon, 29 Mar 2021 08:26:03 GMT
- Title: Federated Semi-Supervised Learning with Inter-Client Consistency &
Disjoint Learning
- Authors: Wonyong Jeong, Jaehong Yoon, Eunho Yang, and Sung Ju Hwang
- Abstract summary: We study two essential scenarios of Federated Semi-Supervised Learning (FSSL) based on the location of the labeled data.
We propose a novel method to tackle the problems, which we refer to as Federated Matching (FedMatch)
- Score: 78.88007892742438
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While existing federated learning approaches mostly require that clients have
fully-labeled data to train on, in realistic settings, data obtained at the
client-side often comes without any accompanying labels. Such deficiency of
labels may result from either high labeling cost, or difficulty of annotation
due to the requirement of expert knowledge. Thus the private data at each
client may be either partly labeled, or completely unlabeled with labeled data
being available only at the server, which leads us to a new practical federated
learning problem, namely Federated Semi-Supervised Learning (FSSL). In this
work, we study two essential scenarios of FSSL based on the location of the
labeled data. The first scenario considers a conventional case where clients
have both labeled and unlabeled data (labels-at-client), and the second
scenario considers a more challenging case, where the labeled data is only
available at the server (labels-at-server). We then propose a novel method to
tackle the problems, which we refer to as Federated Matching (FedMatch).
FedMatch improves upon naive combinations of federated learning and
semi-supervised learning approaches with a new inter-client consistency loss
and decomposition of the parameters for disjoint learning on labeled and
unlabeled data. Through extensive experimental validation of our method in the
two different scenarios, we show that our method outperforms both local
semi-supervised learning and baselines which naively combine federated learning
with semi-supervised learning. The code is available at
https://github.com/wyjeong/FedMatch.
Related papers
- (FL)$^2$: Overcoming Few Labels in Federated Semi-Supervised Learning [4.803231218533992]
Federated Learning (FL) is a distributed machine learning framework that trains accurate global models while preserving clients' privacy-sensitive data.
Most FL approaches assume that clients possess labeled data, which is often not the case in practice.
We propose $(FL)2$, a robust training method for unlabeled clients using sharpness-aware consistency regularization.
arXiv Detail & Related papers (2024-10-30T17:15:02Z) - Federated Learning with Only Positive Labels by Exploring Label Correlations [78.59613150221597]
Federated learning aims to collaboratively learn a model by using the data from multiple users under privacy constraints.
In this paper, we study the multi-label classification problem under the federated learning setting.
We propose a novel and generic method termed Federated Averaging by exploring Label Correlations (FedALC)
arXiv Detail & Related papers (2024-04-24T02:22:50Z) - FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness
for Semi-Supervised Learning [73.13448439554497]
Semi-Supervised Learning (SSL) has been an effective way to leverage abundant unlabeled data with extremely scarce labeled data.
Most SSL methods are commonly based on instance-wise consistency between different data transformations.
We propose FlatMatch which minimizes a cross-sharpness measure to ensure consistent learning performance between the two datasets.
arXiv Detail & Related papers (2023-10-25T06:57:59Z) - JointMatch: A Unified Approach for Diverse and Collaborative
Pseudo-Labeling to Semi-Supervised Text Classification [65.268245109828]
Semi-supervised text classification (SSTC) has gained increasing attention due to its ability to leverage unlabeled data.
Existing approaches based on pseudo-labeling suffer from the issues of pseudo-label bias and error accumulation.
We propose JointMatch, a holistic approach for SSTC that addresses these challenges by unifying ideas from recent semi-supervised learning.
arXiv Detail & Related papers (2023-10-23T05:43:35Z) - Federated Semi-Supervised Learning with Annotation Heterogeneity [57.12560313403097]
We propose a novel framework called Heterogeneously Annotated Semi-Supervised LEarning (HASSLE)
It is a dual-model framework with two models trained separately on labeled and unlabeled data.
The dual models can implicitly learn from both types of data across different clients, although each dual model is only trained locally on a single type of data.
arXiv Detail & Related papers (2023-03-04T16:04:49Z) - FedIL: Federated Incremental Learning from Decentralized Unlabeled Data
with Convergence Analysis [23.70951896315126]
This work considers the server with a small labeled dataset and intends to use unlabeled data in multiple clients for semi-supervised learning.
We propose a new framework with a generalized model, Federated Incremental Learning (FedIL), to address the problem of how to utilize labeled data in the server and unlabeled data in clients separately.
arXiv Detail & Related papers (2023-02-23T07:12:12Z) - Federated Semi-Supervised Learning with Class Distribution Mismatch [34.46190258291223]
Federated semi-supervised learning (Fed-SSL) is an attractive solution for fully utilizing both labeled and unlabeled data.
We introduce two proper regularization terms that can effectively alleviate the class distribution mismatch problem in Fed-SSL.
We leverage the variance reduction and normalized averaging techniques to develop a novel Fed-SSL algorithm.
arXiv Detail & Related papers (2021-10-29T14:18:20Z) - SemiFed: Semi-supervised Federated Learning with Consistency and
Pseudo-Labeling [14.737638416823772]
Federated learning enables multiple clients, such as mobile phones and organizations, to collaboratively learn a shared model for prediction.
In this work, we focus on a new scenario for cross-silo federated learning, where data samples of each client are partially labeled.
We propose a new framework dubbed SemiFed that unifies two dominant approaches for semi-supervised learning: consistency regularization and pseudo-labeling.
arXiv Detail & Related papers (2021-08-21T01:14:27Z) - Federated Semi-supervised Medical Image Classification via Inter-client
Relation Matching [58.26619456972598]
Federated learning (FL) has emerged with increasing popularity to collaborate distributed medical institutions for training deep networks.
This paper studies a practical yet challenging FL problem, named textitFederated Semi-supervised Learning (FSSL)
We present a novel approach for this problem, which improves over traditional consistency regularization mechanism with a new inter-client relation matching scheme.
arXiv Detail & Related papers (2021-06-16T07:58:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.