Graph Generative Model for Benchmarking Graph Neural Networks
- URL: http://arxiv.org/abs/2207.04396v4
- Date: Fri, 9 Jun 2023 11:52:42 GMT
- Title: Graph Generative Model for Benchmarking Graph Neural Networks
- Authors: Minji Yoon, Yue Wu, John Palowitch, Bryan Perozzi, Ruslan
Salakhutdinov
- Abstract summary: We introduce a novel graph generative model that learns and reproduces the distribution of real-world graphs in a privacy-controlled way.
Our model can successfully generate privacy-controlled, synthetic substitutes of large-scale real-world graphs that can be effectively used to benchmark GNN models.
- Score: 73.11514658000547
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As the field of Graph Neural Networks (GNN) continues to grow, it experiences
a corresponding increase in the need for large, real-world datasets to train
and test new GNN models on challenging, realistic problems. Unfortunately, such
graph datasets are often generated from online, highly privacy-restricted
ecosystems, which makes research and development on these datasets hard, if not
impossible. This greatly reduces the amount of benchmark graphs available to
researchers, causing the field to rely only on a handful of publicly-available
datasets. To address this problem, we introduce a novel graph generative model,
Computation Graph Transformer (CGT) that learns and reproduces the distribution
of real-world graphs in a privacy-controlled way. More specifically, CGT (1)
generates effective benchmark graphs on which GNNs show similar task
performance as on the source graphs, (2) scales to process large-scale graphs,
(3) incorporates off-the-shelf privacy modules to guarantee end-user privacy of
the generated graph. Extensive experiments across a vast body of graph
generative models show that only our model can successfully generate
privacy-controlled, synthetic substitutes of large-scale real-world graphs that
can be effectively used to benchmark GNN models.
Related papers
- Data Augmentation in Graph Neural Networks: The Role of Generated Synthetic Graphs [0.24999074238880487]
This study explores using generated graphs for data augmentation.
It compares the performance of combining generated graphs with real graphs, and examining the effect of different quantities of generated graphs on graph classification tasks.
Our results introduce a new approach to graph data augmentation, ensuring consistent labels and enhancing classification performance.
arXiv Detail & Related papers (2024-07-20T06:05:26Z) - Spectral Greedy Coresets for Graph Neural Networks [61.24300262316091]
The ubiquity of large-scale graphs in node-classification tasks hinders the real-world applications of Graph Neural Networks (GNNs)
This paper studies graph coresets for GNNs and avoids the interdependence issue by selecting ego-graphs based on their spectral embeddings.
Our spectral greedy graph coreset (SGGC) scales to graphs with millions of nodes, obviates the need for model pre-training, and applies to low-homophily graphs.
arXiv Detail & Related papers (2024-05-27T17:52:12Z) - Examining the Effects of Degree Distribution and Homophily in Graph
Learning Models [19.060710813929354]
GraphWorld is a solution which generates diverse populations of synthetic graphs for benchmarking any GNN task.
Despite its success, the SBM imposed fundamental limitations on the kinds of graph structure GraphWorld could create.
In this work we examine how two additional synthetic graph generators can improve GraphWorld's evaluation.
arXiv Detail & Related papers (2023-07-17T22:35:46Z) - Edge Directionality Improves Learning on Heterophilic Graphs [42.5099159786891]
We introduce Directed Graph Neural Network (Dir-GNN), a novel framework for deep learning on directed graphs.
Dir-GNN can be used to extend any Message Passing Neural Network (MPNN) to account for edge directionality information.
We prove that Dir-GNN matches the expressivity of the Directed Weisfeiler-Lehman test, exceeding that of conventional MPNNs.
arXiv Detail & Related papers (2023-05-17T18:06:43Z) - Model Inversion Attacks against Graph Neural Networks [65.35955643325038]
We study model inversion attacks against Graph Neural Networks (GNNs)
In this paper, we present GraphMI to infer the private training graph data.
Our experimental results show that such defenses are not sufficiently effective and call for more advanced defenses against privacy attacks.
arXiv Detail & Related papers (2022-09-16T09:13:43Z) - SizeShiftReg: a Regularization Method for Improving Size-Generalization
in Graph Neural Networks [5.008597638379227]
Graph neural networks (GNNs) have become the de facto model of choice for graph classification.
We propose a regularization strategy that can be applied to any GNN to improve its generalization capabilities without requiring access to the test data.
Our regularization is based on the idea of simulating a shift in the size of the training graphs using coarsening techniques.
arXiv Detail & Related papers (2022-07-16T09:50:45Z) - GraphWorld: Fake Graphs Bring Real Insights for GNNs [4.856486822139849]
GraphWorld allows a user to efficiently generate a world with millions of statistically diverse datasets.
We present insights from GraphWorld experiments regarding the performance characteristics of tens of thousands of GNN models over millions of benchmark datasets.
arXiv Detail & Related papers (2022-02-28T22:00:02Z) - GraphMI: Extracting Private Graph Data from Graph Neural Networks [59.05178231559796]
We present textbfGraph textbfModel textbfInversion attack (GraphMI), which aims to extract private graph data of the training graph by inverting GNN.
Specifically, we propose a projected gradient module to tackle the discreteness of graph edges while preserving the sparsity and smoothness of graph features.
We design a graph auto-encoder module to efficiently exploit graph topology, node attributes, and target model parameters for edge inference.
arXiv Detail & Related papers (2021-06-05T07:07:52Z) - Learning to Drop: Robust Graph Neural Network via Topological Denoising [50.81722989898142]
We propose PTDNet, a parameterized topological denoising network, to improve the robustness and generalization performance of Graph Neural Networks (GNNs)
PTDNet prunes task-irrelevant edges by penalizing the number of edges in the sparsified graph with parameterized networks.
We show that PTDNet can improve the performance of GNNs significantly and the performance gain becomes larger for more noisy datasets.
arXiv Detail & Related papers (2020-11-13T18:53:21Z) - GPT-GNN: Generative Pre-Training of Graph Neural Networks [93.35945182085948]
Graph neural networks (GNNs) have been demonstrated to be powerful in modeling graph-structured data.
We present the GPT-GNN framework to initialize GNNs by generative pre-training.
We show that GPT-GNN significantly outperforms state-of-the-art GNN models without pre-training by up to 9.1% across various downstream tasks.
arXiv Detail & Related papers (2020-06-27T20:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.