Improving Semi-supervised Federated Learning by Reducing the Gradient
Diversity of Models
- URL: http://arxiv.org/abs/2008.11364v2
- Date: Tue, 9 Mar 2021 04:03:44 GMT
- Title: Improving Semi-supervised Federated Learning by Reducing the Gradient
Diversity of Models
- Authors: Zhengming Zhang, Yaoqing Yang, Zhewei Yao, Yujun Yan, Joseph E.
Gonzalez, Michael W. Mahoney
- Abstract summary: Federated learning (FL) is a promising way to use the computing power of mobile devices while maintaining privacy of users.
We show that a critical issue that affects the test accuracy is the large gradient diversity of the models from different users.
We propose a novel grouping-based model averaging method to replace the FedAvg averaging method.
- Score: 67.66144604972052
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning (FL) is a promising way to use the computing power of
mobile devices while maintaining the privacy of users. Current work in FL,
however, makes the unrealistic assumption that the users have ground-truth
labels on their devices, while also assuming that the server has neither data
nor labels. In this work, we consider the more realistic scenario where the
users have only unlabeled data, while the server has some labeled data, and
where the amount of labeled data is smaller than the amount of unlabeled data.
We call this learning problem semi-supervised federated learning (SSFL). For
SSFL, we demonstrate that a critical issue that affects the test accuracy is
the large gradient diversity of the models from different users. Based on this,
we investigate several design choices. First, we find that the so-called
consistency regularization loss (CRL), which is widely used in semi-supervised
learning, performs reasonably well but has large gradient diversity. Second, we
find that Batch Normalization (BN) increases gradient diversity. Replacing BN
with the recently-proposed Group Normalization (GN) can reduce gradient
diversity and improve test accuracy. Third, we show that CRL combined with GN
still has a large gradient diversity when the number of users is large. Based
on these results, we propose a novel grouping-based model averaging method to
replace the FedAvg averaging method. Overall, our grouping-based averaging,
combined with GN and CRL, achieves better test accuracy than not just a
contemporary paper on SSFL in the same settings (>10\%), but also four
supervised FL algorithms.
Related papers
- Towards Realistic Long-tailed Semi-supervised Learning in an Open World [0.0]
We construct a more emphRealistic Open-world Long-tailed Semi-supervised Learning (textbfROLSSL) setting where there is no premise on the distribution relationships between known and novel categories.
Under the proposed ROLSSL setting, we propose a simple yet potentially effective solution called dual-stage logit adjustments.
Experiments on datasets such as CIFAR100 and ImageNet100 have demonstrated performance improvements of up to 50.1%.
arXiv Detail & Related papers (2024-05-23T12:53:50Z) - Communication Efficient ConFederated Learning: An Event-Triggered SAGA
Approach [67.27031215756121]
Federated learning (FL) is a machine learning paradigm that targets model training without gathering the local data over various data sources.
Standard FL, which employs a single server, can only support a limited number of users, leading to degraded learning capability.
In this work, we consider a multi-server FL framework, referred to as emphConfederated Learning (CFL) in order to accommodate a larger number of users.
arXiv Detail & Related papers (2024-02-28T03:27:10Z) - Exploiting Label Skews in Federated Learning with Model Concatenation [39.38427550571378]
Federated Learning (FL) has emerged as a promising solution to perform deep learning on different data owners without exchanging raw data.
Among different non-IID types, label skews have been challenging and common in image classification and other tasks.
We propose FedConcat, a simple and effective approach that degrades these local models as the base of the global model.
arXiv Detail & Related papers (2023-12-11T10:44:52Z) - Rethinking Data Heterogeneity in Federated Learning: Introducing a New
Notion and Standard Benchmarks [65.34113135080105]
We show that not only the issue of data heterogeneity in current setups is not necessarily a problem but also in fact it can be beneficial for the FL participants.
Our observations are intuitive.
Our code is available at https://github.com/MMorafah/FL-SC-NIID.
arXiv Detail & Related papers (2022-09-30T17:15:19Z) - StatMix: Data augmentation method that relies on image statistics in
federated learning [0.27528170226206433]
StatMix is an augmentation approach that uses image statistics to improve results of FL scenario(s)
In all FL experiments, application of StatMix improves the average accuracy, compared to the baseline training (with no use of StatMix)
arXiv Detail & Related papers (2022-07-08T19:02:41Z) - Local Learning Matters: Rethinking Data Heterogeneity in Federated
Learning [61.488646649045215]
Federated learning (FL) is a promising strategy for performing privacy-preserving, distributed learning with a network of clients (i.e., edge devices)
arXiv Detail & Related papers (2021-11-28T19:03:39Z) - Multi-Center Federated Learning [62.32725938999433]
Federated learning (FL) can protect data privacy in distributed learning.
It merely collects local gradients from users without access to their data.
We propose a novel multi-center aggregation mechanism.
arXiv Detail & Related papers (2021-08-19T12:20:31Z) - FedSemi: An Adaptive Federated Semi-Supervised Learning Framework [23.90642104477983]
Federated learning (FL) has emerged as an effective technique to co-training machine learning models without actually sharing data and leaking privacy.
Most existing FL methods focus on the supervised setting and ignore the utilization of unlabeled data.
We propose FedSemi, a novel, adaptive, and general framework, which firstly introduces the consistency regularization into FL using a teacher-student model.
arXiv Detail & Related papers (2020-12-06T15:46:04Z) - Multi-Center Federated Learning [62.57229809407692]
This paper proposes a novel multi-center aggregation mechanism for federated learning.
It learns multiple global models from the non-IID user data and simultaneously derives the optimal matching between users and centers.
Our experimental results on benchmark datasets show that our method outperforms several popular federated learning methods.
arXiv Detail & Related papers (2020-05-03T09:14:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.