Distribution-Controlled Client Selection to Improve Federated Learning Strategies
- URL: http://arxiv.org/abs/2509.20877v1
- Date: Thu, 25 Sep 2025 08:07:13 GMT
- Title: Distribution-Controlled Client Selection to Improve Federated Learning Strategies
- Authors: Christoph Düsing, Philipp Cimiano,
- Abstract summary: Federated learning (FL) is a distributed learning paradigm that allows multiple clients to jointly train a shared model.<n>The presence of data imbalance among clients is a thread to the success of FL, as it causes the performance of the shared model to decrease.<n>We propose an extension to existing FL strategies, which selects active clients that best align the current label distribution with one of two target distributions.
- Score: 4.254099382808598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning (FL) is a distributed learning paradigm that allows multiple clients to jointly train a shared model while maintaining data privacy. Despite its great potential for domains with strict data privacy requirements, the presence of data imbalance among clients is a thread to the success of FL, as it causes the performance of the shared model to decrease. To address this, various studies have proposed enhancements to existing FL strategies, particularly through client selection methods that mitigate the detrimental effects of data imbalance. In this paper, we propose an extension to existing FL strategies, which selects active clients that best align the current label distribution with one of two target distributions, namely a balanced distribution or the federations combined label distribution. Subsequently, we empirically verify the improvements through our distribution-controlled client selection on three common FL strategies and two datasets. Our results show that while aligning the label distribution with a balanced distribution yields the greatest improvements facing local imbalance, alignment with the federation's combined label distribution is superior for global imbalance.
Related papers
- Balancing Client Participation in Federated Learning Using AoI [3.2999744336237384]
Federated Learning (FL) offers a decentralized framework that preserves data privacy while enabling collaborative model training across distributed clients.<n>This paper proposes an Age of Information (AoI)-based client selection policy that addresses these challenges by minimizing load imbalance through controlled selection intervals.
arXiv Detail & Related papers (2025-05-08T09:55:28Z) - FedDiverse: Tackling Data Heterogeneity in Federated Learning with Diversity-Driven Client Selection [21.93193941981852]
Federated Learning (FL) enables decentralized training of machine learning models on distributed data.<n>In real-world FL settings, client data is often non-identically distributed and imbalanced.<n>We propose FEDDIVERSE, a novel client selection algorithm in FL which is designed to manage and leverage data heterogeneity across clients.
arXiv Detail & Related papers (2025-04-15T14:20:42Z) - Disentangling data distribution for Federated Learning [20.524108508314107]
Federated Learning (FL) facilitates collaborative training of a global model whose performance is boosted by private data owned by distributed clients.<n>Yet the wide applicability of FL is hindered by entanglement of data distributions across different clients.<n>This paper demonstrates for the first time that by disentangling data distributions FL can in principle achieve efficiencies comparable to those of distributed systems.
arXiv Detail & Related papers (2024-10-16T13:10:04Z) - FedImpro: Measuring and Improving Client Update in Federated Learning [77.68805026788836]
Federated Learning (FL) models often experience client drift caused by heterogeneous data.
We present an alternative perspective on client drift and aim to mitigate it by generating improved local models.
arXiv Detail & Related papers (2024-02-10T18:14:57Z) - Overcome Modal Bias in Multi-modal Federated Learning via Balanced Modality Selection [19.284989473603627]
We propose a novel Balanced Modality Selection framework for multi-modal learning (MFL)
We show that local training with a certain single modality may contribute more to the global model than training with all local modalities.
Our experiments on audio-visual, colored-gray, and front-back datasets showcase the superiority of BMSFed over baselines.
arXiv Detail & Related papers (2023-12-31T05:37:27Z) - PS-FedGAN: An Efficient Federated Learning Framework Based on Partially
Shared Generative Adversarial Networks For Data Privacy [56.347786940414935]
Federated Learning (FL) has emerged as an effective learning paradigm for distributed computation.
This work proposes a novel FL framework that requires only partial GAN model sharing.
Named as PS-FedGAN, this new framework enhances the GAN releasing and training mechanism to address heterogeneous data distributions.
arXiv Detail & Related papers (2023-05-19T05:39:40Z) - FedABC: Targeting Fair Competition in Personalized Federated Learning [76.9646903596757]
Federated learning aims to collaboratively train models without accessing their client's local private data.
We propose a novel and generic PFL framework termed Federated Averaging via Binary Classification, dubbed FedABC.
In particular, we adopt the one-vs-all'' training strategy in each client to alleviate the unfair competition between classes.
arXiv Detail & Related papers (2023-02-15T03:42:59Z) - FL Games: A Federated Learning Framework for Distribution Shifts [71.98708418753786]
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
We propose FL GAMES, a game-theoretic framework for federated learning that learns causal features that are invariant across clients.
arXiv Detail & Related papers (2022-10-31T22:59:03Z) - Fed-CBS: A Heterogeneity-Aware Client Sampling Mechanism for Federated
Learning via Class-Imbalance Reduction [76.26710990597498]
We show that the class-imbalance of the grouped data from randomly selected clients can lead to significant performance degradation.
Based on our key observation, we design an efficient client sampling mechanism, i.e., Federated Class-balanced Sampling (Fed-CBS)
In particular, we propose a measure of class-imbalance and then employ homomorphic encryption to derive this measure in a privacy-preserving way.
arXiv Detail & Related papers (2022-09-30T05:42:56Z) - FL Games: A federated learning framework for distribution shifts [71.98708418753786]
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
We propose FL Games, a game-theoretic framework for federated learning for learning causal features that are invariant across clients.
arXiv Detail & Related papers (2022-05-23T07:51:45Z) - Dubhe: Towards Data Unbiasedness with Homomorphic Encryption in
Federated Learning Client Selection [16.975086164684882]
Federated learning (FL) is a distributed machine learning paradigm that allows clients to collaboratively train a model over their own local data.
We mathematically demonstrate the cause of performance degradation in FL and examine the performance of FL over various datasets.
We propose a pluggable system-level client selection method named Dubhe, which allows clients to proactively participate in training, preserving their privacy with the assistance of HE.
arXiv Detail & Related papers (2021-09-08T13:00:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.