FedCorr: Multi-Stage Federated Learning for Label Noise Correction
- URL: http://arxiv.org/abs/2204.04677v1
- Date: Sun, 10 Apr 2022 12:51:18 GMT
- Title: FedCorr: Multi-Stage Federated Learning for Label Noise Correction
- Authors: Jingyi Xu, Zihan Chen, Tony Q.S. Quek, Kai Fong Ernest Chong
- Abstract summary: Federated learning (FL) is a privacy-preserving distributed learning paradigm that enables clients to jointly train a global model.
We propose $textttFedCorr$, a general multi-stage framework to tackle heterogeneous label noise in FL.
Experiments conducted on CIFAR-10/100 with federated synthetic label noise, and on a real-world noisy dataset, Clothing1M, demonstrate that $textttFedCorr$ is robust to label noise.
- Score: 80.9366438220228
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning (FL) is a privacy-preserving distributed learning paradigm
that enables clients to jointly train a global model. In real-world FL
implementations, client data could have label noise, and different clients
could have vastly different label noise levels. Although there exist methods in
centralized learning for tackling label noise, such methods do not perform well
on heterogeneous label noise in FL settings, due to the typically smaller sizes
of client datasets and data privacy requirements in FL. In this paper, we
propose $\texttt{FedCorr}$, a general multi-stage framework to tackle
heterogeneous label noise in FL, without making any assumptions on the noise
models of local clients, while still maintaining client data privacy. In
particular, (1) $\texttt{FedCorr}$ dynamically identifies noisy clients by
exploiting the dimensionalities of the model prediction subspaces independently
measured on all clients, and then identifies incorrect labels on noisy clients
based on per-sample losses. To deal with data heterogeneity and to increase
training stability, we propose an adaptive local proximal regularization term
that is based on estimated local noise levels. (2) We further finetune the
global model on identified clean clients and correct the noisy labels for the
remaining noisy clients after finetuning. (3) Finally, we apply the usual
training on all clients to make full use of all local data. Experiments
conducted on CIFAR-10/100 with federated synthetic label noise, and on a
real-world noisy dataset, Clothing1M, demonstrate that $\texttt{FedCorr}$ is
robust to label noise and substantially outperforms the state-of-the-art
methods at multiple noise levels.
Related papers
- Federated Learning Client Pruning for Noisy Labels [6.30126491637621]
Federated Learning (FL) enables collaborative model training across decentralized edge devices.
This paper introduces ClipFL, a novel framework addressing noisy labels from a fresh perspective.
It identifies and excludes noisy clients based on their performance on a clean validation dataset.
arXiv Detail & Related papers (2024-11-11T21:46:34Z) - FedFixer: Mitigating Heterogeneous Label Noise in Federated Learning [15.382625332503125]
Federated Learning (FL) heavily depends on label quality for its performance.
The high loss incurred by client-specific samples in heterogeneous label noise poses challenges for distinguishing between client-specific and noisy label samples.
We propose FedFixer, where the personalized model is introduced to cooperate with the global model to effectively select clean client-specific samples.
arXiv Detail & Related papers (2024-03-25T09:24:05Z) - Federated Learning with Extremely Noisy Clients via Negative
Distillation [70.13920804879312]
Federated learning (FL) has shown remarkable success in cooperatively training deep models, while struggling with noisy labels.
We propose a novel approach, called negative distillation (FedNed) to leverage models trained on noisy clients.
FedNed first identifies noisy clients and employs rather than discards the noisy clients in a knowledge distillation manner.
arXiv Detail & Related papers (2023-12-20T01:59:48Z) - FedDiv: Collaborative Noise Filtering for Federated Learning with Noisy
Labels [99.70895640578816]
Federated learning with noisy labels (F-LNL) aims at seeking an optimal server model via collaborative distributed learning.
We present FedDiv to tackle the challenges of F-LNL. Specifically, we propose a global noise filter called Federated Noise Filter.
arXiv Detail & Related papers (2023-12-19T15:46:47Z) - Neighborhood Collective Estimation for Noisy Label Identification and
Correction [92.20697827784426]
Learning with noisy labels (LNL) aims at designing strategies to improve model performance and generalization by mitigating the effects of model overfitting to noisy labels.
Recent advances employ the predicted label distributions of individual samples to perform noise verification and noisy label correction, easily giving rise to confirmation bias.
We propose Neighborhood Collective Estimation, in which the predictive reliability of a candidate sample is re-estimated by contrasting it against its feature-space nearest neighbors.
arXiv Detail & Related papers (2022-08-05T14:47:22Z) - FedNoiL: A Simple Two-Level Sampling Method for Federated Learning with
Noisy Labels [49.47228898303909]
Federated learning (FL) aims at training a global model on the server side while the training data are collected and located at the local devices.
Local training on noisy labels can easily result in overfitting to noisy labels, which is devastating to the global model through aggregation.
We develop a simple two-level sampling method "FedNoiL" that selects clients for more robust global aggregation on the server.
arXiv Detail & Related papers (2022-05-20T12:06:39Z) - Federated Noisy Client Learning [105.00756772827066]
Federated learning (FL) collaboratively aggregates a shared global model depending on multiple local clients.
Standard FL methods ignore the noisy client issue, which may harm the overall performance of the aggregated model.
We propose Federated Noisy Client Learning (Fed-NCL), which is a plug-and-play algorithm and contains two main components.
arXiv Detail & Related papers (2021-06-24T11:09:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.