Micro and Macro Level Graph Modeling for Graph Variational Auto-Encoders
- URL: http://arxiv.org/abs/2210.16844v1
- Date: Sun, 30 Oct 2022 13:45:21 GMT
- Title: Micro and Macro Level Graph Modeling for Graph Variational Auto-Encoders
- Authors: Kiarash Zahirnia, Oliver Schulte, Parmis Naddaf, Ke Li
- Abstract summary: This paper proposes a new multi-level framework that jointly models node-level properties and graph-level statistics.
We introduce a new micro-macro training objective for graph generation that combines node-level and graph-level losses.
Our experiments show that adding micro-macro modeling to the GraphVAE model improves graph quality scores up to 2 orders of magnitude on five benchmark datasets.
- Score: 16.302222204710276
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative models for graph data are an important research topic in machine
learning. Graph data comprise two levels that are typically analyzed
separately: node-level properties such as the existence of a link between a
pair of nodes, and global aggregate graph-level statistics, such as motif
counts. This paper proposes a new multi-level framework that jointly models
node-level properties and graph-level statistics, as mutually reinforcing
sources of information. We introduce a new micro-macro training objective for
graph generation that combines node-level and graph-level losses. We utilize
the micro-macro objective to improve graph generation with a GraphVAE, a
well-established model based on graph-level latent variables, that provides
fast training and generation time for medium-sized graphs. Our experiments show
that adding micro-macro modeling to the GraphVAE model improves graph quality
scores up to 2 orders of magnitude on five benchmark datasets, while
maintaining the GraphVAE generation speed advantage.
Related papers
- Graph Transformer GANs with Graph Masked Modeling for Architectural
Layout Generation [153.92387500677023]
We present a novel graph Transformer generative adversarial network (GTGAN) to learn effective graph node relations.
The proposed graph Transformer encoder combines graph convolutions and self-attentions in a Transformer to model both local and global interactions.
We also propose a novel self-guided pre-training method for graph representation learning.
arXiv Detail & Related papers (2024-01-15T14:36:38Z) - GraphMaker: Can Diffusion Models Generate Large Attributed Graphs? [7.330479039715941]
Large-scale graphs with node attributes are increasingly common in various real-world applications.
Traditional graph generation methods are limited in their capacity to handle these complex structures.
This paper introduces a novel diffusion model, GraphMaker, specifically designed for generating large attributed graphs.
arXiv Detail & Related papers (2023-10-20T22:12:46Z) - GDM: Dual Mixup for Graph Classification with Limited Supervision [27.8982897698616]
Graph Neural Networks (GNNs) require a large number of labeled graph samples to obtain good performance on the graph classification task.
The performance of GNNs degrades significantly as the number of labeled graph samples decreases.
We propose a novel mixup-based graph augmentation method to generate new labeled graph samples.
arXiv Detail & Related papers (2023-09-18T20:17:10Z) - Structure-free Graph Condensation: From Large-scale Graphs to Condensed
Graph-free Data [91.27527985415007]
Existing graph condensation methods rely on the joint optimization of nodes and structures in the condensed graph.
We advocate a new Structure-Free Graph Condensation paradigm, named SFGC, to distill a large-scale graph into a small-scale graph node set.
arXiv Detail & Related papers (2023-06-05T07:53:52Z) - Towards Graph Self-Supervised Learning with Contrastive Adjusted Zooming [48.99614465020678]
We introduce a novel self-supervised graph representation learning algorithm via Graph Contrastive Adjusted Zooming.
This mechanism enables G-Zoom to explore and extract self-supervision signals from a graph from multiple scales.
We have conducted extensive experiments on real-world datasets, and the results demonstrate that our proposed model outperforms state-of-the-art methods consistently.
arXiv Detail & Related papers (2021-11-20T22:45:53Z) - Generating the Graph Gestalt: Kernel-Regularized Graph Representation
Learning [47.506013386710954]
A complete scientific understanding of graph data should address both global and local structure.
We propose a joint model for both as complementary objectives in a graph VAE framework.
Our experiments demonstrate a significant improvement in the realism of the generated graph structures, typically by 1-2 orders of magnitude of graph structure metrics.
arXiv Detail & Related papers (2021-06-29T10:48:28Z) - Dirichlet Graph Variational Autoencoder [65.94744123832338]
We present Dirichlet Graph Variational Autoencoder (DGVAE) with graph cluster memberships as latent factors.
Motivated by the low pass characteristics in balanced graph cut, we propose a new variant of GNN named Heatts to encode the input graph into cluster memberships.
arXiv Detail & Related papers (2020-10-09T07:35:26Z) - Multilevel Graph Matching Networks for Deep Graph Similarity Learning [79.3213351477689]
We propose a multi-level graph matching network (MGMN) framework for computing the graph similarity between any pair of graph-structured objects.
To compensate for the lack of standard benchmark datasets, we have created and collected a set of datasets for both the graph-graph classification and graph-graph regression tasks.
Comprehensive experiments demonstrate that MGMN consistently outperforms state-of-the-art baseline models on both the graph-graph classification and graph-graph regression tasks.
arXiv Detail & Related papers (2020-07-08T19:48:19Z) - Adaptive Graph Auto-Encoder for General Data Clustering [90.8576971748142]
Graph-based clustering plays an important role in the clustering area.
Recent studies about graph convolution neural networks have achieved impressive success on graph type data.
We propose a graph auto-encoder for general data clustering, which constructs the graph adaptively according to the generative perspective of graphs.
arXiv Detail & Related papers (2020-02-20T10:11:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.