Addressing Data Heterogeneity in Decentralized Learning via Topological
Pre-processing
- URL: http://arxiv.org/abs/2212.08743v1
- Date: Fri, 16 Dec 2022 22:46:38 GMT
- Title: Addressing Data Heterogeneity in Decentralized Learning via Topological
Pre-processing
- Authors: Waqwoya Abebe, Ali Jannesari
- Abstract summary: We show the advantages of constructing a proxy-based locally heterogeneous DL topology to enhance convergence and maintain data privacy.
We propose a novel peer clumping strategy to efficiently cluster peers before arranging them in a final training graph.
- Score: 0.9645196221785693
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, local peer topology has been shown to influence the overall
convergence of decentralized learning (DL) graphs in the presence of data
heterogeneity. In this paper, we demonstrate the advantages of constructing a
proxy-based locally heterogeneous DL topology to enhance convergence and
maintain data privacy. In particular, we propose a novel peer clumping strategy
to efficiently cluster peers before arranging them in a final training graph.
By showing how locally heterogeneous graphs outperform locally homogeneous
graphs of similar size and from the same global data distribution, we present a
strong case for topological pre-processing. Moreover, we demonstrate the
scalability of our approach by showing how the proposed topological
pre-processing overhead remains small in large graphs while the performance
gains get even more pronounced. Furthermore, we show the robustness of our
approach in the presence of network partitions.
Related papers
- Self-Supervised Contrastive Graph Clustering Network via Structural Information Fusion [15.293684479404092]
We propose a novel deep graph clustering method called CGCN.
Our approach introduces contrastive signals and deep structural information into the pre-training process.
Our method has been experimentally validated on multiple real-world graph datasets.
arXiv Detail & Related papers (2024-08-08T09:49:26Z) - Revealing Decurve Flows for Generalized Graph Propagation [108.80758541147418]
This study addresses the limitations of the traditional analysis of message-passing, central to graph learning, by defining em textbfgeneralized propagation with directed and weighted graphs.
We include a preliminary exploration of learned propagation patterns in datasets, a first in the field.
arXiv Detail & Related papers (2024-02-13T14:13:17Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - Interpolation-based Correlation Reduction Network for Semi-Supervised
Graph Learning [49.94816548023729]
We propose a novel graph contrastive learning method, termed Interpolation-based Correlation Reduction Network (ICRN)
In our method, we improve the discriminative capability of the latent feature by enlarging the margin of decision boundaries.
By combining the two settings, we extract rich supervision information from both the abundant unlabeled nodes and the rare yet valuable labeled nodes for discnative representation learning.
arXiv Detail & Related papers (2022-06-06T14:26:34Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Data-heterogeneity-aware Mixing for Decentralized Learning [63.83913592085953]
We characterize the dependence of convergence on the relationship between the mixing weights of the graph and the data heterogeneity across nodes.
We propose a metric that quantifies the ability of a graph to mix the current gradients.
Motivated by our analysis, we propose an approach that periodically and efficiently optimize the metric.
arXiv Detail & Related papers (2022-04-13T15:54:35Z) - Yes, Topology Matters in Decentralized Optimization: Refined Convergence
and Topology Learning under Heterogeneous Data [0.0]
We revisit the analysis of Decentralized Gradient Descent algorithm (D-SGD), a popular decentralized learning algorithm, under data heterogeneity.
We exhibit the key role played by a new quantity, that we call neighborhood heterogeneity, on the convergence rate of D-SGD.
By coupling the topology and the heterogeneity of the agents' distributions, our analysis sheds light on the poorly understood interplay between these two concepts in decentralized learning.
arXiv Detail & Related papers (2022-04-09T11:29:12Z) - A Graph Data Augmentation Strategy with Entropy Preserving [11.886325179121226]
We introduce a novel graph entropy definition as a quantitative index to evaluate feature information among a graph.
Under considerations of preserving graph entropy, we propose an effective strategy to generate training data using a perturbed mechanism.
Our proposed approach significantly enhances the robustness and generalization ability of GCNs during the training process.
arXiv Detail & Related papers (2021-07-13T12:58:32Z) - Contrastive and Generative Graph Convolutional Networks for Graph-based
Semi-Supervised Learning [64.98816284854067]
Graph-based Semi-Supervised Learning (SSL) aims to transfer the labels of a handful of labeled data to the remaining massive unlabeled data via a graph.
A novel GCN-based SSL algorithm is presented in this paper to enrich the supervision signals by utilizing both data similarities and graph structure.
arXiv Detail & Related papers (2020-09-15T13:59:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.