GDM: Dual Mixup for Graph Classification with Limited Supervision
- URL: http://arxiv.org/abs/2309.10134v1
- Date: Mon, 18 Sep 2023 20:17:10 GMT
- Title: GDM: Dual Mixup for Graph Classification with Limited Supervision
- Authors: Abdullah Alchihabi and Yuhong Guo
- Abstract summary: Graph Neural Networks (GNNs) require a large number of labeled graph samples to obtain good performance on the graph classification task.
The performance of GNNs degrades significantly as the number of labeled graph samples decreases.
We propose a novel mixup-based graph augmentation method to generate new labeled graph samples.
- Score: 27.8982897698616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Neural Networks (GNNs) require a large number of labeled graph samples
to obtain good performance on the graph classification task. The performance of
GNNs degrades significantly as the number of labeled graph samples decreases.
To reduce the annotation cost, it is therefore important to develop graph
augmentation methods that can generate new graph instances to increase the size
and diversity of the limited set of available labeled graph samples. In this
work, we propose a novel mixup-based graph augmentation method, Graph Dual
Mixup (GDM), that leverages both functional and structural information of the
graph instances to generate new labeled graph samples. GDM employs a graph
structural auto-encoder to learn structural embeddings of the graph samples,
and then applies mixup to the structural information of the graphs in the
learned structural embedding space and generates new graph structures from the
mixup structural embeddings. As for the functional information, GDM applies
mixup directly to the input node features of the graph samples to generate
functional node feature information for new mixup graph instances. Jointly, the
generated input node features and graph structures yield new graph samples
which can supplement the set of original labeled graphs. Furthermore, we
propose two novel Balanced Graph Sampling methods to enhance the balanced
difficulty and diversity for the generated graph samples. Experimental results
on the benchmark datasets demonstrate that our proposed method substantially
outperforms the state-of-the-art graph augmentation methods when the labeled
graphs are scarce.
Related papers
- InstructG2I: Synthesizing Images from Multimodal Attributed Graphs [50.852150521561676]
We propose a graph context-conditioned diffusion model called InstructG2I.
InstructG2I first exploits the graph structure and multimodal information to conduct informative neighbor sampling.
A Graph-QFormer encoder adaptively encodes the graph nodes into an auxiliary set of graph prompts to guide the denoising process.
arXiv Detail & Related papers (2024-10-09T17:56:15Z) - Data Augmentation in Graph Neural Networks: The Role of Generated Synthetic Graphs [0.24999074238880487]
This study explores using generated graphs for data augmentation.
It compares the performance of combining generated graphs with real graphs, and examining the effect of different quantities of generated graphs on graph classification tasks.
Our results introduce a new approach to graph data augmentation, ensuring consistent labels and enhancing classification performance.
arXiv Detail & Related papers (2024-07-20T06:05:26Z) - Greener GRASS: Enhancing GNNs with Encoding, Rewiring, and Attention [12.409982249220812]
We introduce Graph Attention with Structures (GRASS), a novel GNN architecture, to enhance graph relative attention.
GRASS rewires the input graph by superimposing a random regular graph to achieve long-range information propagation.
It also employs a novel additive attention mechanism tailored for graph-structured data.
arXiv Detail & Related papers (2024-07-08T06:21:56Z) - GraphAlign: Pretraining One Graph Neural Network on Multiple Graphs via Feature Alignment [30.56443056293688]
Graph self-supervised learning (SSL) holds considerable promise for mining and learning with graph-structured data.
In this work, we aim to pretrain one graph neural network (GNN) on a varied collection of graphs endowed with rich node features.
We present a general GraphAlign method that can be seamlessly integrated into the existing graph SSL framework.
arXiv Detail & Related papers (2024-06-05T05:22:32Z) - GraphMaker: Can Diffusion Models Generate Large Attributed Graphs? [7.330479039715941]
Large-scale graphs with node attributes are increasingly common in various real-world applications.
Traditional graph generation methods are limited in their capacity to handle these complex structures.
This paper introduces a novel diffusion model, GraphMaker, specifically designed for generating large attributed graphs.
arXiv Detail & Related papers (2023-10-20T22:12:46Z) - Graph Mixup with Soft Alignments [49.61520432554505]
We study graph data augmentation by mixup, which has been used successfully on images.
We propose S-Mixup, a simple yet effective mixup method for graph classification by soft alignments.
arXiv Detail & Related papers (2023-06-11T22:04:28Z) - Structure-free Graph Condensation: From Large-scale Graphs to Condensed
Graph-free Data [91.27527985415007]
Existing graph condensation methods rely on the joint optimization of nodes and structures in the condensed graph.
We advocate a new Structure-Free Graph Condensation paradigm, named SFGC, to distill a large-scale graph into a small-scale graph node set.
arXiv Detail & Related papers (2023-06-05T07:53:52Z) - Graph Condensation via Receptive Field Distribution Matching [61.71711656856704]
This paper focuses on creating a small graph to represent the original graph, so that GNNs trained on the size-reduced graph can make accurate predictions.
We view the original graph as a distribution of receptive fields and aim to synthesize a small graph whose receptive fields share a similar distribution.
arXiv Detail & Related papers (2022-06-28T02:10:05Z) - Multilevel Graph Matching Networks for Deep Graph Similarity Learning [79.3213351477689]
We propose a multi-level graph matching network (MGMN) framework for computing the graph similarity between any pair of graph-structured objects.
To compensate for the lack of standard benchmark datasets, we have created and collected a set of datasets for both the graph-graph classification and graph-graph regression tasks.
Comprehensive experiments demonstrate that MGMN consistently outperforms state-of-the-art baseline models on both the graph-graph classification and graph-graph regression tasks.
arXiv Detail & Related papers (2020-07-08T19:48:19Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.