Improving Graph Generation by Restricting Graph Bandwidth
- URL: http://arxiv.org/abs/2301.10857v2
- Date: Tue, 30 May 2023 22:56:34 GMT
- Title: Improving Graph Generation by Restricting Graph Bandwidth
- Authors: Nathaniel Diamant, Alex M. Tseng, Kangway V. Chuang, Tommaso
Biancalani, Gabriele Scalia
- Abstract summary: Deep graph generative modeling has proven capable of learning the distribution of complex, multi-scale structures characterizing real-world graphs.
One of the main limitations of existing methods is their large output space.
We propose a novel approach that significantly reduces the output space of existing graph generative models.
- Score: 0.7874708385247353
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep graph generative modeling has proven capable of learning the
distribution of complex, multi-scale structures characterizing real-world
graphs. However, one of the main limitations of existing methods is their large
output space, which limits generation scalability and hinders accurate modeling
of the underlying distribution. To overcome these limitations, we propose a
novel approach that significantly reduces the output space of existing graph
generative models. Specifically, starting from the observation that many
real-world graphs have low graph bandwidth, we restrict graph bandwidth during
training and generation. Our strategy improves both generation scalability and
quality without increasing architectural complexity or reducing expressiveness.
Our approach is compatible with existing graph generative methods, and we
describe its application to both autoregressive and one-shot models. We
extensively validate our strategy on synthetic and real datasets, including
molecular graphs. Our experiments show that, in addition to improving
generation efficiency, our approach consistently improves generation quality
and reconstruction accuracy. The implementation is made available.
Related papers
- Dynamic and Textual Graph Generation Via Large-Scale LLM-based Agent Simulation [70.60461609393779]
GraphAgent-Generator (GAG) is a novel simulation-based framework for dynamic graph generation.
Our framework effectively replicates seven macro-level structural characteristics in established network science theories.
It supports generating graphs with up to nearly 100,000 nodes or 10 million edges, with a minimum speed-up of 90.4%.
arXiv Detail & Related papers (2024-10-13T12:57:08Z) - AnyGraph: Graph Foundation Model in the Wild [16.313146933922752]
Graph foundation models offer the potential to learn robust, generalizable representations from graph data.
In this work, we investigate a unified graph model, AnyGraph, designed to handle key challenges.
Our experiments on diverse 38 graph datasets have demonstrated the strong zero-shot learning performance of AnyGraph.
arXiv Detail & Related papers (2024-08-20T09:57:13Z) - GLAD: Improving Latent Graph Generative Modeling with Simple Quantization [44.86731507203063]
We propose a simple, yet effective discrete latent graph diffusion generative model.
Our model, namely GLAD, not only overcomes the drawbacks of existing latent approaches, but also alleviates inherent issues present in diffusion methods applied on the graph space.
We validate our generative model on the molecular benchmark datasets, on which it demonstrates competitive performance compared with the state-of-the-art baselines.
arXiv Detail & Related papers (2024-03-25T15:53:32Z) - GraphRCG: Self-Conditioned Graph Generation [78.69810678803248]
We propose a novel self-conditioned graph generation framework designed to explicitly model graph distributions.
Our framework demonstrates superior performance over existing state-of-the-art graph generation methods in terms of graph quality and fidelity to training data.
arXiv Detail & Related papers (2024-03-02T02:28:20Z) - A Topology-aware Graph Coarsening Framework for Continual Graph Learning [8.136809136959302]
Continual learning on graphs tackles the problem of training a graph neural network (GNN) where graph data arrive in a streaming fashion.
Traditional continual learning strategies such as Experience Replay can be adapted to streaming graphs.
We propose TA$mathbbCO$, a (t)opology-(a)ware graph (co)arsening and (co)ntinual learning framework.
arXiv Detail & Related papers (2024-01-05T22:22:13Z) - GraphMaker: Can Diffusion Models Generate Large Attributed Graphs? [7.330479039715941]
Large-scale graphs with node attributes are increasingly common in various real-world applications.
Traditional graph generation methods are limited in their capacity to handle these complex structures.
This paper introduces a novel diffusion model, GraphMaker, specifically designed for generating large attributed graphs.
arXiv Detail & Related papers (2023-10-20T22:12:46Z) - HiGen: Hierarchical Graph Generative Networks [2.3931689873603603]
Most real-world graphs exhibit a hierarchical structure, which is often overlooked by existing graph generation methods.
We propose a novel graph generative network that captures the hierarchical nature of graphs and successively generates the graph sub-structures in a coarse-to-fine fashion.
This modular approach enables scalable graph generation for large and complex graphs.
arXiv Detail & Related papers (2023-05-30T18:04:12Z) - FairGen: Towards Fair Graph Generation [76.34239875010381]
We propose a fairness-aware graph generative model named FairGen.
Our model jointly trains a label-informed graph generation module and a fair representation learning module.
Experimental results on seven real-world data sets, including web-based graphs, demonstrate that FairGen obtains performance on par with state-of-the-art graph generative models.
arXiv Detail & Related papers (2023-03-30T23:30:42Z) - Graph Generation with Diffusion Mixture [57.78958552860948]
Generation of graphs is a major challenge for real-world tasks that require understanding the complex nature of their non-Euclidean structures.
We propose a generative framework that models the topology of graphs by explicitly learning the final graph structures of the diffusion process.
arXiv Detail & Related papers (2023-02-07T17:07:46Z) - Towards Graph Self-Supervised Learning with Contrastive Adjusted Zooming [48.99614465020678]
We introduce a novel self-supervised graph representation learning algorithm via Graph Contrastive Adjusted Zooming.
This mechanism enables G-Zoom to explore and extract self-supervision signals from a graph from multiple scales.
We have conducted extensive experiments on real-world datasets, and the results demonstrate that our proposed model outperforms state-of-the-art methods consistently.
arXiv Detail & Related papers (2021-11-20T22:45:53Z) - Adaptive Graph Auto-Encoder for General Data Clustering [90.8576971748142]
Graph-based clustering plays an important role in the clustering area.
Recent studies about graph convolution neural networks have achieved impressive success on graph type data.
We propose a graph auto-encoder for general data clustering, which constructs the graph adaptively according to the generative perspective of graphs.
arXiv Detail & Related papers (2020-02-20T10:11:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.