K-Core Decomposition on Super Large Graphs with Limited Resources
- URL: http://arxiv.org/abs/2112.14840v1
- Date: Sun, 26 Dec 2021 04:34:11 GMT
- Title: K-Core Decomposition on Super Large Graphs with Limited Resources
- Authors: Shicheng Gao, Jie Xu, Xiaosen Li, Fangcheng Fu, Wentao Zhang, Wen
Ouyang, Yangyu Tao, Bin Cui
- Abstract summary: Recent years have seen rapid growth in the scale of the graph, especially in industrial settings.
Applying K-core decomposition on large graphs has attracted more and more attention from academics and the industry.
We propose a divide-and-conquer strategy on top of the distributed K-core decomposition algorithm.
- Score: 17.71064869466004
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: K-core decomposition is a commonly used metric to analyze graph structure or
study the relative importance of nodes in complex graphs. Recent years have
seen rapid growth in the scale of the graph, especially in industrial settings.
For example, our industrial partner runs popular social applications with
billions of users and is able to gather a rich set of user data. As a result,
applying K-core decomposition on large graphs has attracted more and more
attention from academics and the industry. A simple but effective method to
deal with large graphs is to train them in the distributed settings, and some
distributed K-core decomposition algorithms are also proposed. Despite their
effectiveness, we experimentally and theoretically observe that these
algorithms consume too many resources and become unstable on super-large-scale
graphs, especially when the given resources are limited. In this paper, we deal
with those super-large-scale graphs and propose a divide-and-conquer strategy
on top of the distributed K-core decomposition algorithm. We evaluate our
approach on three large graphs. The experimental results show that the
consumption of resources can be significantly reduced, and the calculation on
large-scale graphs becomes more stable than the existing methods. For example,
the distributed K-core decomposition algorithm can scale to a large graph with
136 billion edges without losing correctness with our divide-and-conquer
technique.
Related papers
- Expander Hierarchies for Normalized Cuts on Graphs [3.3385430106181184]
We introduce first practically efficient algorithm for computing expander decompositions and their hierarchies.
Our experiments on a variety of large graphs show that our expander-based algorithm outperforms state-of-the-art solvers for normalized cut.
arXiv Detail & Related papers (2024-06-20T08:50:57Z) - Spectral Greedy Coresets for Graph Neural Networks [61.24300262316091]
The ubiquity of large-scale graphs in node-classification tasks hinders the real-world applications of Graph Neural Networks (GNNs)
This paper studies graph coresets for GNNs and avoids the interdependence issue by selecting ego-graphs based on their spectral embeddings.
Our spectral greedy graph coreset (SGGC) scales to graphs with millions of nodes, obviates the need for model pre-training, and applies to low-homophily graphs.
arXiv Detail & Related papers (2024-05-27T17:52:12Z) - Improved Exact and Heuristic Algorithms for Maximum Weight Clique [1.2074552857379273]
We propose improved exact and labeling algorithms for solving the maximum weight clique problem.
Our algorithms interleave successful techniques with novel data reduction rules that use local graph structure.
We evaluate our algorithms on a range of synthetic and real-world graphs, and find that they outperform the current state of the art on most inputs.
arXiv Detail & Related papers (2023-02-01T14:02:06Z) - Graph Condensation via Receptive Field Distribution Matching [61.71711656856704]
This paper focuses on creating a small graph to represent the original graph, so that GNNs trained on the size-reduced graph can make accurate predictions.
We view the original graph as a distribution of receptive fields and aim to synthesize a small graph whose receptive fields share a similar distribution.
arXiv Detail & Related papers (2022-06-28T02:10:05Z) - GraphCoCo: Graph Complementary Contrastive Learning [65.89743197355722]
Graph Contrastive Learning (GCL) has shown promising performance in graph representation learning (GRL) without the supervision of manual annotations.
This paper proposes an effective graph complementary contrastive learning approach named GraphCoCo to tackle the above issue.
arXiv Detail & Related papers (2022-03-24T02:58:36Z) - Scaling R-GCN Training with Graph Summarization [71.06855946732296]
Training of Relation Graph Convolutional Networks (R-GCN) does not scale well with the size of the graph.
In this work, we experiment with the use of graph summarization techniques to compress the graph.
We obtain reasonable results on the AIFB, MUTAG and AM datasets.
arXiv Detail & Related papers (2022-03-05T00:28:43Z) - Graph Coarsening with Neural Networks [8.407217618651536]
We propose a framework for measuring the quality of coarsening algorithm and show that depending on the goal, we need to carefully choose the Laplace operator on the coarse graph.
Motivated by the observation that the current choice of edge weight for the coarse graph may be sub-optimal, we parametrize the weight assignment map with graph neural networks and train it to improve the coarsening quality in an unsupervised way.
arXiv Detail & Related papers (2021-02-02T06:50:07Z) - LCS Graph Kernel Based on Wasserstein Distance in Longest Common
Subsequence Metric Space [22.215852332444904]
We propose a graph kernel to compute more comprehensive similarity between paths and walks.
We also combine it with optimal transport theory to extract more in-depth features of graphs.
Our proposed method outperforms many state-of-the-art graph kernel methods.
arXiv Detail & Related papers (2020-12-07T11:59:14Z) - Multilayer Clustered Graph Learning [66.94201299553336]
We use contrastive loss as a data fidelity term, in order to properly aggregate the observed layers into a representative graph.
Experiments show that our method leads to a clustered clusters w.r.t.
We learn a clustering algorithm for solving clustering problems.
arXiv Detail & Related papers (2020-10-29T09:58:02Z) - Understanding Coarsening for Embedding Large-Scale Graphs [3.6739949215165164]
Proper analysis of graphs with Machine Learning (ML) algorithms has the potential to yield far-reaching insights into many areas of research and industry.
The irregular structure of graph data constitutes an obstacle for running ML tasks on graphs.
We analyze the impact of the coarsening quality on the embedding performance both in terms of speed and accuracy.
arXiv Detail & Related papers (2020-09-10T15:06:33Z) - Block-Approximated Exponential Random Graphs [77.4792558024487]
An important challenge in the field of exponential random graphs (ERGs) is the fitting of non-trivial ERGs on large graphs.
We propose an approximative framework to such non-trivial ERGs that result in dyadic independence (i.e., edge independent) distributions.
Our methods are scalable to sparse graphs consisting of millions of nodes.
arXiv Detail & Related papers (2020-02-14T11:42:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.