AugWard: Augmentation-Aware Representation Learning for Accurate Graph Classification
- URL: http://arxiv.org/abs/2503.21105v1
- Date: Thu, 27 Mar 2025 02:58:28 GMT
- Title: AugWard: Augmentation-Aware Representation Learning for Accurate Graph Classification
- Authors: Minjun Kim, Jaehyeon Choi, SeungJoo Lee, Jinhong Jung, U Kang,
- Abstract summary: AugWard is a graph representation learning framework that considers the diversity introduced by graph augmentation.<n>AugWard applies augmentation-aware training to predict the graph distance between the augmented graph and its original one.<n>Results show that AugWard gives the state-of-the-art performance in supervised, semi-supervised graph classification, and transfer learning.
- Score: 16.7104207718009
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: How can we accurately classify graphs? Graph classification is a pivotal task in data mining with applications in social network analysis, web analysis, drug discovery, molecular property prediction, etc. Graph neural networks have achieved the state-of-the-art performance in graph classification, but they consistently struggle with overfitting. To mitigate overfitting, researchers have introduced various representation learning methods utilizing graph augmentation. However, existing methods rely on simplistic use of graph augmentation, which loses augmentation-induced differences and limits the expressiveness of representations. In this paper, we propose AugWard (Augmentation-Aware Training with Graph Distance and Consistency Regularization), a novel graph representation learning framework that carefully considers the diversity introduced by graph augmentation. AugWard applies augmentation-aware training to predict the graph distance between the augmented graph and its original one, aligning the representation difference directly with graph distance at both feature and structure levels. Furthermore, AugWard employs consistency regularization to encourage the classifier to handle richer representations. Experimental results show that AugWard gives the state-of-the-art performance in supervised, semi-supervised graph classification, and transfer learning.
Related papers
- Motif-Consistent Counterfactuals with Adversarial Refinement for Graph-Level Anomaly Detection [30.618065157205507]
We propose a novel approach, Motif-consistent Counterfactuals with Adversarial Refinement (MotifCAR) for graph-level anomaly detection.
The model combines the motif of one graph, the core subgraph containing the identification (category) information, and the contextual subgraph of another graph to produce a raw counterfactual graph.
MotifCAR can generate high-quality counterfactual graphs.
arXiv Detail & Related papers (2024-07-18T08:04:57Z) - Spectral Augmentations for Graph Contrastive Learning [50.149996923976836]
Contrastive learning has emerged as a premier method for learning representations with or without supervision.
Recent studies have shown its utility in graph representation learning for pre-training.
We propose a set of well-motivated graph transformation operations to provide a bank of candidates when constructing augmentations for a graph contrastive objective.
arXiv Detail & Related papers (2023-02-06T16:26:29Z) - State of the Art and Potentialities of Graph-level Learning [54.68482109186052]
Graph-level learning has been applied to many tasks including comparison, regression, classification, and more.
Traditional approaches to learning a set of graphs rely on hand-crafted features, such as substructures.
Deep learning has helped graph-level learning adapt to the growing scale of graphs by extracting features automatically and encoding graphs into low-dimensional representations.
arXiv Detail & Related papers (2023-01-14T09:15:49Z) - CGMN: A Contrastive Graph Matching Network for Self-Supervised Graph
Similarity Learning [65.1042892570989]
We propose a contrastive graph matching network (CGMN) for self-supervised graph similarity learning.
We employ two strategies, namely cross-view interaction and cross-graph interaction, for effective node representation learning.
We transform node representations into graph-level representations via pooling operations for graph similarity computation.
arXiv Detail & Related papers (2022-05-30T13:20:26Z) - Model-Agnostic Augmentation for Accurate Graph Classification [19.824105919844495]
Graph augmentation is an essential strategy to improve the performance of graph-based tasks.
In this work, we introduce five desired properties for effective augmentation.
Our experiments on social networks and molecular graphs show that NodeSam and SubMix outperform existing approaches in graph classification.
arXiv Detail & Related papers (2022-02-21T10:37:53Z) - Graph Self-supervised Learning with Accurate Discrepancy Learning [64.69095775258164]
We propose a framework that aims to learn the exact discrepancy between the original and the perturbed graphs, coined as Discrepancy-based Self-supervised LeArning (D-SLA)
We validate our method on various graph-related downstream tasks, including molecular property prediction, protein function prediction, and link prediction tasks, on which our model largely outperforms relevant baselines.
arXiv Detail & Related papers (2022-02-07T08:04:59Z) - Bootstrapping Informative Graph Augmentation via A Meta Learning
Approach [21.814940639910358]
In graph contrastive learning, benchmark methods apply various graph augmentation approaches.
Most of the augmentation methods are non-learnable, which causes the issue of generating unbeneficial augmented graphs.
We motivate our method to generate augmented graph by a learnable graph augmenter, called MEta Graph Augmentation (MEGA)
arXiv Detail & Related papers (2022-01-11T07:15:13Z) - Unbiased Graph Embedding with Biased Graph Observations [52.82841737832561]
We propose a principled new way for obtaining unbiased representations by learning from an underlying bias-free graph.
Based on this new perspective, we propose two complementary methods for uncovering such an underlying graph.
arXiv Detail & Related papers (2021-10-26T18:44:37Z) - Graph Contrastive Learning with Augmentations [109.23158429991298]
We propose a graph contrastive learning (GraphCL) framework for learning unsupervised representations of graph data.
We show that our framework can produce graph representations of similar or better generalizability, transferrability, and robustness compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-10-22T20:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.