FedNoiL: A Simple Two-Level Sampling Method for Federated Learning with
Noisy Labels
- URL: http://arxiv.org/abs/2205.10110v1
- Date: Fri, 20 May 2022 12:06:39 GMT
- Title: FedNoiL: A Simple Two-Level Sampling Method for Federated Learning with
Noisy Labels
- Authors: Zhuowei Wang, Tianyi Zhou, Guodong Long, Bo Han, Jing Jiang
- Abstract summary: Federated learning (FL) aims at training a global model on the server side while the training data are collected and located at the local devices.
Local training on noisy labels can easily result in overfitting to noisy labels, which is devastating to the global model through aggregation.
We develop a simple two-level sampling method "FedNoiL" that selects clients for more robust global aggregation on the server.
- Score: 49.47228898303909
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning (FL) aims at training a global model on the server side
while the training data are collected and located at the local devices. Hence,
the labels in practice are usually annotated by clients of varying expertise or
criteria and thus contain different amounts of noises. Local training on noisy
labels can easily result in overfitting to noisy labels, which is devastating
to the global model through aggregation. Although recent robust FL methods take
malicious clients into account, they have not addressed local noisy labels on
each device and the impact to the global model. In this paper, we develop a
simple two-level sampling method "FedNoiL" that (1) selects clients for more
robust global aggregation on the server; and (2) selects clean labels and
correct pseudo-labels at the client end for more robust local training. The
sampling probabilities are built upon clean label detection by the global
model. Moreover, we investigate different schedules changing the local epochs
between aggregations over the course of FL, which notably improves the
communication and computation efficiency in noisy label setting. In experiments
with homogeneous/heterogeneous data distributions and noise ratios, we observed
that direct combinations of SOTA FL methods with SOTA noisy-label learning
methods can easily fail but our method consistently achieves better and robust
performance.
Related papers
- FedFixer: Mitigating Heterogeneous Label Noise in Federated Learning [15.382625332503125]
Federated Learning (FL) heavily depends on label quality for its performance.
The high loss incurred by client-specific samples in heterogeneous label noise poses challenges for distinguishing between client-specific and noisy label samples.
We propose FedFixer, where the personalized model is introduced to cooperate with the global model to effectively select clean client-specific samples.
arXiv Detail & Related papers (2024-03-25T09:24:05Z) - FedDiv: Collaborative Noise Filtering for Federated Learning with Noisy
Labels [99.70895640578816]
Federated learning with noisy labels (F-LNL) aims at seeking an optimal server model via collaborative distributed learning.
We present FedDiv to tackle the challenges of F-LNL. Specifically, we propose a global noise filter called Federated Noise Filter.
arXiv Detail & Related papers (2023-12-19T15:46:47Z) - Federated Learning with Instance-Dependent Noisy Label [6.093214616626228]
FedBeat aims to build a global statistically consistent classifier using the IDN transition matrix (IDNTM)
Experiments conducted on CIFAR-10 and SVHN verify that the proposed method significantly outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-12-16T05:08:02Z) - Rethinking Client Drift in Federated Learning: A Logit Perspective [125.35844582366441]
Federated Learning (FL) enables multiple clients to collaboratively learn in a distributed way, allowing for privacy protection.
We find that the difference in logits between the local and global models increases as the model is continuously updated.
We propose a new algorithm, named FedCSD, a Class prototype Similarity Distillation in a federated framework to align the local and global models.
arXiv Detail & Related papers (2023-08-20T04:41:01Z) - Learning Cautiously in Federated Learning with Noisy and Heterogeneous
Clients [4.782145666637457]
Federated learning (FL) is a distributed framework for collaboratively training with privacy guarantees.
In real-world scenarios, clients may have Non-IID data (local class imbalance) with poor annotation quality (label noise)
We propose FedCNI without using an additional clean proxy dataset.
It includes a noise-resilient local solver and a robust global aggregator.
arXiv Detail & Related papers (2023-04-06T06:47:14Z) - Neighborhood Collective Estimation for Noisy Label Identification and
Correction [92.20697827784426]
Learning with noisy labels (LNL) aims at designing strategies to improve model performance and generalization by mitigating the effects of model overfitting to noisy labels.
Recent advances employ the predicted label distributions of individual samples to perform noise verification and noisy label correction, easily giving rise to confirmation bias.
We propose Neighborhood Collective Estimation, in which the predictive reliability of a candidate sample is re-estimated by contrasting it against its feature-space nearest neighbors.
arXiv Detail & Related papers (2022-08-05T14:47:22Z) - FedCorr: Multi-Stage Federated Learning for Label Noise Correction [80.9366438220228]
Federated learning (FL) is a privacy-preserving distributed learning paradigm that enables clients to jointly train a global model.
We propose $textttFedCorr$, a general multi-stage framework to tackle heterogeneous label noise in FL.
Experiments conducted on CIFAR-10/100 with federated synthetic label noise, and on a real-world noisy dataset, Clothing1M, demonstrate that $textttFedCorr$ is robust to label noise.
arXiv Detail & Related papers (2022-04-10T12:51:18Z) - Federated Noisy Client Learning [105.00756772827066]
Federated learning (FL) collaboratively aggregates a shared global model depending on multiple local clients.
Standard FL methods ignore the noisy client issue, which may harm the overall performance of the aggregated model.
We propose Federated Noisy Client Learning (Fed-NCL), which is a plug-and-play algorithm and contains two main components.
arXiv Detail & Related papers (2021-06-24T11:09:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.