Walk for Learning: A Random Walk Approach for Federated Learning from
Heterogeneous Data
- URL: http://arxiv.org/abs/2206.00737v1
- Date: Wed, 1 Jun 2022 19:53:24 GMT
- Title: Walk for Learning: A Random Walk Approach for Federated Learning from
Heterogeneous Data
- Authors: Ghadir Ayache, Venkat Dassari, Salim El Rouayheb
- Abstract summary: We focus on Federated Learning (FL) as a canonical application.
One of the main challenges of FL is the communication bottleneck between the nodes and the parameter server.
We present an adaptive random walk learning algorithm.
- Score: 17.978941229970886
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the problem of a Parameter Server (PS) that wishes to learn a
model that fits data distributed on the nodes of a graph. We focus on Federated
Learning (FL) as a canonical application. One of the main challenges of FL is
the communication bottleneck between the nodes and the parameter server. A
popular solution in the literature is to allow each node to do several local
updates on the model in each iteration before sending it back to the PS. While
this mitigates the communication bottleneck, the statistical heterogeneity of
the data owned by the different nodes has proven to delay convergence and bias
the model. In this work, we study random walk (RW) learning algorithms for
tackling the communication and data heterogeneity problems. The main idea is to
leverage available direct connections among the nodes themselves, which are
typically "cheaper" than the communication to the PS. In a random walk, the
model is thought of as a "baton" that is passed from a node to one of its
neighbors after being updated in each iteration. The challenge in designing the
RW is the data heterogeneity and the uncertainty about the data distributions.
Ideally, we would want to visit more often nodes that hold more informative
data. We cast this problem as a sleeping multi-armed bandit (MAB) to design a
near-optimal node sampling strategy that achieves variance-reduced gradient
estimates and approaches sub-linearly the optimal sampling strategy. Based on
this framework, we present an adaptive random walk learning algorithm. We
provide theoretical guarantees on its convergence. Our numerical results
validate our theoretical findings and show that our algorithm outperforms
existing random walk algorithms.
Related papers
- The Heterophilic Snowflake Hypothesis: Training and Empowering GNNs for Heterophilic Graphs [59.03660013787925]
We introduce the Heterophily Snowflake Hypothesis and provide an effective solution to guide and facilitate research on heterophilic graphs.
Our observations show that our framework acts as a versatile operator for diverse tasks.
It can be integrated into various GNN frameworks, boosting performance in-depth and offering an explainable approach to choosing the optimal network depth.
arXiv Detail & Related papers (2024-06-18T12:16:00Z) - Communication-Efficient Decentralized Federated Learning via One-Bit
Compressive Sensing [52.402550431781805]
Decentralized federated learning (DFL) has gained popularity due to its practicality across various applications.
Compared to the centralized version, training a shared model among a large number of nodes in DFL is more challenging.
We develop a novel algorithm based on the framework of the inexact alternating direction method (iADM)
arXiv Detail & Related papers (2023-08-31T12:22:40Z) - Neural-prior stochastic block model [0.0]
We propose to model the communities as being determined by the node attributes rather than the opposite.
We propose an algorithm, stemming from statistical physics, based on a combination of belief propagation and approximate message passing.
The proposed model and algorithm can be used as a benchmark for both theory and algorithms.
arXiv Detail & Related papers (2023-03-17T14:14:54Z) - Faster Adaptive Federated Learning [84.38913517122619]
Federated learning has attracted increasing attention with the emergence of distributed data.
In this paper, we propose an efficient adaptive algorithm (i.e., FAFED) based on momentum-based variance reduced technique in cross-silo FL.
arXiv Detail & Related papers (2022-12-02T05:07:50Z) - QuTE: decentralized multiple testing on sensor networks with false
discovery rate control [130.7122910646076]
This paper designs methods for decentralized multiple hypothesis testing on graphs equipped with provable guarantees on the false discovery rate (FDR)
We consider the setting where distinct agents reside on the nodes of an undirected graph, and each agent possesses p-values corresponding to one or more hypotheses local to its node.
Each agent must individually decide whether to reject one or more of its local hypotheses by only communicating with its neighbors, with the joint aim that the global FDR over the entire graph must be controlled at a predefined level.
arXiv Detail & Related papers (2022-10-09T19:48:39Z) - STEM: A Stochastic Two-Sided Momentum Algorithm Achieving Near-Optimal
Sample and Communication Complexities for Federated Learning [58.6792963686231]
Federated Learning (FL) refers to the paradigm where multiple worker nodes (WNs) build a joint model by using local data.
It is not clear how to choose the WNs' minimum update directions, the first minibatch sizes, and the local update frequency.
We show that there is a trade-off curve between local update frequencies and local mini sizes, on which the above complexities can be maintained.
arXiv Detail & Related papers (2021-06-19T06:13:45Z) - Private Weighted Random Walk Stochastic Gradient Descent [21.23973736418492]
We consider a decentralized learning setting in which data is distributed over nodes in a graph.
We propose two algorithms based on two types of random walks that achieve, in a decentralized way, uniform sampling and importance sampling of the data.
arXiv Detail & Related papers (2020-09-03T16:52:13Z) - FedPD: A Federated Learning Framework with Optimal Rates and Adaptivity
to Non-IID Data [59.50904660420082]
Federated Learning (FL) has become a popular paradigm for learning from distributed data.
To effectively utilize data at different devices without moving them to the cloud, algorithms such as the Federated Averaging (FedAvg) have adopted a "computation then aggregation" (CTA) model.
arXiv Detail & Related papers (2020-05-22T23:07:42Z) - PushNet: Efficient and Adaptive Neural Message Passing [1.9121961872220468]
Message passing neural networks have recently evolved into a state-of-the-art approach to representation learning on graphs.
Existing methods perform synchronous message passing along all edges in multiple subsequent rounds.
We consider a novel asynchronous message passing approach where information is pushed only along the most relevant edges until convergence.
arXiv Detail & Related papers (2020-03-04T18:15:30Z) - Fully Asynchronous Policy Evaluation in Distributed Reinforcement
Learning over Networks [14.636457985379746]
This paper proposes a emphfully asynchronous scheme for the policy evaluation problem of distributed reinforcement learning (DisRL) over directed peer-to-peer networks.
Without waiting for any other node of the network, each node can locally update its value function at any time by using (possibly delayed) information from its neighbors.
arXiv Detail & Related papers (2020-03-01T08:12:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.