Game of Gradients: Mitigating Irrelevant Clients in Federated Learning
- URL: http://arxiv.org/abs/2110.12257v1
- Date: Sat, 23 Oct 2021 16:34:42 GMT
- Title: Game of Gradients: Mitigating Irrelevant Clients in Federated Learning
- Authors: Lokesh Nagalapatti, Ramasuri Narayanam
- Abstract summary: Federated learning (FL) deals with multiple clients participating in collaborative training of a machine learning model under the orchestration of a central server.
In this setup, each client's data is private to itself and is not transferable to other clients or the server.
We refer to these problems as Federated Relevant Client Selection (FRCS)
- Score: 3.2095659532757916
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The paradigm of Federated learning (FL) deals with multiple clients
participating in collaborative training of a machine learning model under the
orchestration of a central server. In this setup, each client's data is private
to itself and is not transferable to other clients or the server. Though FL
paradigm has received significant interest recently from the research
community, the problem of selecting the relevant clients w.r.t. the central
server's learning objective is under-explored. We refer to these problems as
Federated Relevant Client Selection (FRCS). Because the server doesn't have
explicit control over the nature of data possessed by each client, the problem
of selecting relevant clients is significantly complex in FL settings. In this
paper, we resolve important and related FRCS problems viz., selecting clients
with relevant data, detecting clients that possess data relevant to a
particular target label, and rectifying corrupted data samples of individual
clients. We follow a principled approach to address the above FRCS problems and
develop a new federated learning method using the Shapley value concept from
cooperative game theory. Towards this end, we propose a cooperative game
involving the gradients shared by the clients. Using this game, we compute
Shapley values of clients and then present Shapley value based Federated
Averaging (S-FedAvg) algorithm that empowers the server to select relevant
clients with high probability. S-FedAvg turns out to be critical in designing
specific algorithms to address the FRCS problems. We finally conduct a thorough
empirical analysis on image classification and speech recognition tasks to show
the superior performance of S-FedAvg than the baselines in the context of
supervised federated learning settings.
Related papers
- Fine-Tuning Personalization in Federated Learning to Mitigate Adversarial Clients [8.773068878015856]
Federated learning (FL) is an appealing paradigm that allows a group of machines (a.k.a. clients) to learn collectively while keeping their data local.
We consider an FL setting where some clients can be adversarial, and we derive conditions under which full collaboration fails.
arXiv Detail & Related papers (2024-09-30T14:31:19Z) - Utilizing Free Clients in Federated Learning for Focused Model
Enhancement [9.370655190768163]
Federated Learning (FL) is a distributed machine learning approach to learn models on decentralized heterogeneous data.
We present FedALIGN (Federated Adaptive Learning with Inclusion of Global Needs) to address this challenge.
arXiv Detail & Related papers (2023-10-06T18:23:40Z) - Towards Instance-adaptive Inference for Federated Learning [80.38701896056828]
Federated learning (FL) is a distributed learning paradigm that enables multiple clients to learn a powerful global model by aggregating local training.
In this paper, we present a novel FL algorithm, i.e., FedIns, to handle intra-client data heterogeneity by enabling instance-adaptive inference in the FL framework.
Our experiments show that our FedIns outperforms state-of-the-art FL algorithms, e.g., a 6.64% improvement against the top-performing method with less than 15% communication cost on Tiny-ImageNet.
arXiv Detail & Related papers (2023-08-11T09:58:47Z) - Graph Federated Learning with Hidden Representation Sharing [33.01999333117515]
Learning on Graphs (LoG) is widely used in multi-client systems when each client has insufficient local data.
Federated Learning (FL) requires models to be trained in a multi-client system and sharing of raw data among clients.
In this work, we first formulate the Graph Federated Learning (GFL) problem that unifies LoG FL in multi-client systems and then propose sharing hidden representation.
arXiv Detail & Related papers (2022-12-23T05:44:27Z) - Scalable Collaborative Learning via Representation Sharing [53.047460465980144]
Federated learning (FL) and Split Learning (SL) are two frameworks that enable collaborative learning while keeping the data private (on device)
In FL, each data holder trains a model locally and releases it to a central server for aggregation.
In SL, the clients must release individual cut-layer activations (smashed data) to the server and wait for its response (during both inference and back propagation).
In this work, we present a novel approach for privacy-preserving machine learning, where the clients collaborate via online knowledge distillation using a contrastive loss.
arXiv Detail & Related papers (2022-11-20T10:49:22Z) - Efficient Distribution Similarity Identification in Clustered Federated
Learning via Principal Angles Between Client Data Subspaces [59.33965805898736]
Clustered learning has been shown to produce promising results by grouping clients into clusters.
Existing FL algorithms are essentially trying to group clients together with similar distributions.
Prior FL algorithms attempt similarities indirectly during training.
arXiv Detail & Related papers (2022-09-21T17:37:54Z) - On the Convergence of Clustered Federated Learning [57.934295064030636]
In a federated learning system, the clients, e.g. mobile devices and organization participants, usually have different personal preferences or behavior patterns.
This paper proposes a novel weighted client-based clustered FL algorithm to leverage the client's group and each client in a unified optimization framework.
arXiv Detail & Related papers (2022-02-13T02:39:19Z) - Federated Multi-Target Domain Adaptation [99.93375364579484]
Federated learning methods enable us to train machine learning models on distributed user data while preserving its privacy.
We consider a more practical scenario where the distributed client data is unlabeled, and a centralized labeled dataset is available on the server.
We propose an effective DualAdapt method to address the new challenges.
arXiv Detail & Related papers (2021-08-17T17:53:05Z) - DeFed: A Principled Decentralized and Privacy-Preserving Federated
Learning Algorithm [10.487593244018933]
Federated learning enables a large number of clients to participate in learning a shared model while maintaining the training data stored in each client.
Here we propose a principled decentralized federated learning algorithm (DeFed), which removes the central client in the classical Federated Averaging (FedAvg) setting.
The proposed algorithm is proven to reach the global minimum with a convergence rate of $O(1/T)$ when the loss function is smooth and strongly convex, where $T$ is the number of iterations in gradient descent.
arXiv Detail & Related papers (2021-07-15T07:39:19Z) - Federated Learning with Taskonomy for Non-IID Data [0.0]
We introduce federated learning with taskonomy.
In a one-off process, the server provides the clients with a pretrained (and fine-tunable) encoder to compress their data into a latent representation, and transmit the signature of their data back to the server.
The server then learns the task-relatedness among clients via manifold learning, and performs a generalization of federated averaging.
arXiv Detail & Related papers (2021-03-29T20:47:45Z) - Blockchain Assisted Decentralized Federated Learning (BLADE-FL):
Performance Analysis and Resource Allocation [119.19061102064497]
We propose a decentralized FL framework by integrating blockchain into FL, namely, blockchain assisted decentralized federated learning (BLADE-FL)
In a round of the proposed BLADE-FL, each client broadcasts its trained model to other clients, competes to generate a block based on the received models, and then aggregates the models from the generated block before its local training of the next round.
We explore the impact of lazy clients on the learning performance of BLADE-FL, and characterize the relationship among the optimal K, the learning parameters, and the proportion of lazy clients.
arXiv Detail & Related papers (2021-01-18T07:19:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.