BayGo: Joint Bayesian Learning and Information-Aware Graph Optimization
- URL: http://arxiv.org/abs/2011.04345v2
- Date: Fri, 19 Feb 2021 19:47:14 GMT
- Title: BayGo: Joint Bayesian Learning and Information-Aware Graph Optimization
- Authors: Tamara Alshammari and Sumudu Samarakoon and Anis Elgabli and Mehdi
Bennis
- Abstract summary: BayGo is a novel fully decentralized joint Bayesian learning and graph optimization framework.
We show that our framework achieves faster convergence and higher accuracy compared to fully-connected and star topology graphs.
- Score: 48.30183416069897
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article deals with the problem of distributed machine learning, in which
agents update their models based on their local datasets, and aggregate the
updated models collaboratively and in a fully decentralized manner. In this
paper, we tackle the problem of information heterogeneity arising in
multi-agent networks where the placement of informative agents plays a crucial
role in the learning dynamics. Specifically, we propose BayGo, a novel fully
decentralized joint Bayesian learning and graph optimization framework with
proven fast convergence over a sparse graph. Under our framework, agents are
able to learn and communicate with the most informative agent to their own
learning. Unlike prior works, our framework assumes no prior knowledge of the
data distribution across agents nor does it assume any knowledge of the true
parameter of the system. The proposed alternating minimization based framework
ensures global connectivity in a fully decentralized way while minimizing the
number of communication links. We theoretically show that by optimizing the
proposed objective function, the estimation error of the posterior probability
distribution decreases exponentially at each iteration. Via extensive
simulations, we show that our framework achieves faster convergence and higher
accuracy compared to fully-connected and star topology graphs.
Related papers
- Self-Supervised Contrastive Graph Clustering Network via Structural Information Fusion [15.293684479404092]
We propose a novel deep graph clustering method called CGCN.
Our approach introduces contrastive signals and deep structural information into the pre-training process.
Our method has been experimentally validated on multiple real-world graph datasets.
arXiv Detail & Related papers (2024-08-08T09:49:26Z) - Distributed Learning over Networks with Graph-Attention-Based
Personalization [49.90052709285814]
We propose a graph-based personalized algorithm (GATTA) for distributed deep learning.
In particular, the personalized model in each agent is composed of a global part and a node-specific part.
By treating each agent as one node in a graph the node-specific parameters as its features, the benefits of the graph attention mechanism can be inherited.
arXiv Detail & Related papers (2023-05-22T13:48:30Z) - Beyond spectral gap (extended): The role of the topology in
decentralized learning [58.48291921602417]
In data-parallel optimization of machine learning models, workers collaborate to improve their estimates of the model.
Current theory does not explain that collaboration enables larger learning rates than training alone.
This paper aims to paint an accurate picture of sparsely-connected distributed optimization.
arXiv Detail & Related papers (2023-01-05T16:53:38Z) - Beyond spectral gap: The role of the topology in decentralized learning [58.48291921602417]
In data-parallel optimization of machine learning models, workers collaborate to improve their estimates of the model.
This paper aims to paint an accurate picture of sparsely-connected distributed optimization when workers share the same data distribution.
Our theory matches empirical observations in deep learning, and accurately describes the relative merits of different graph topologies.
arXiv Detail & Related papers (2022-06-07T08:19:06Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Dif-MAML: Decentralized Multi-Agent Meta-Learning [54.39661018886268]
We propose a cooperative multi-agent meta-learning algorithm, referred to as MAML or Dif-MAML.
We show that the proposed strategy allows a collection of agents to attain agreement at a linear rate and to converge to a stationary point of the aggregate MAML.
Simulation results illustrate the theoretical findings and the superior performance relative to the traditional non-cooperative setting.
arXiv Detail & Related papers (2020-10-06T16:51:09Z) - Jointly Optimizing Dataset Size and Local Updates in Heterogeneous
Mobile Edge Learning [11.191719032853527]
This paper proposes to maximize the accuracy of a distributed machine learning (ML) model trained on learners connected via the resource-constrained wireless edge.
We jointly optimize the number of local/global updates and the task size allocation to minimize the loss while taking into account heterogeneous communication and computation capabilities of each learner.
arXiv Detail & Related papers (2020-06-12T18:19:20Z) - PushNet: Efficient and Adaptive Neural Message Passing [1.9121961872220468]
Message passing neural networks have recently evolved into a state-of-the-art approach to representation learning on graphs.
Existing methods perform synchronous message passing along all edges in multiple subsequent rounds.
We consider a novel asynchronous message passing approach where information is pushed only along the most relevant edges until convergence.
arXiv Detail & Related papers (2020-03-04T18:15:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.