Label-invariant Augmentation for Semi-Supervised Graph Classification
- URL: http://arxiv.org/abs/2205.09802v1
- Date: Thu, 19 May 2022 18:44:02 GMT
- Title: Label-invariant Augmentation for Semi-Supervised Graph Classification
- Authors: Han Yue, Chunhui Zhang, Chuxu Zhang, Hongfu Liu
- Abstract summary: Recently, contrastiveness-based augmentation surges a new climax in the computer vision domain.
Unlike images, it is much more difficult to design reasonable augmentations without changing the nature of graphs.
We propose a label-invariant augmentation for graph-structured data to address this challenge.
- Score: 32.591130704357184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, contrastiveness-based augmentation surges a new climax in the
computer vision domain, where some operations, including rotation, crop, and
flip, combined with dedicated algorithms, dramatically increase the model
generalization and robustness. Following this trend, some pioneering attempts
employ the similar idea to graph data. Nevertheless, unlike images, it is much
more difficult to design reasonable augmentations without changing the nature
of graphs. Although exciting, the current graph contrastive learning does not
achieve as promising performance as visual contrastive learning. We conjecture
the current performance of graph contrastive learning might be limited by the
violation of the label-invariant augmentation assumption. In light of this, we
propose a label-invariant augmentation for graph-structured data to address
this challenge. Different from the node/edge modification and subgraph
extraction, we conduct the augmentation in the representation space and
generate the augmented samples in the most difficult direction while keeping
the label of augmented data the same as the original samples. In the
semi-supervised scenario, we demonstrate our proposed method outperforms the
classical graph neural network based methods and recent graph contrastive
learning on eight benchmark graph-structured data, followed by several in-depth
experiments to further explore the label-invariant augmentation in several
aspects.
Related papers
- Uncovering Capabilities of Model Pruning in Graph Contrastive Learning [0.0]
We reformulate the problem of graph contrastive learning via contrasting different model versions rather than augmented views.
We extensively validate our method on various benchmarks regarding graph classification via unsupervised and transfer learning.
arXiv Detail & Related papers (2024-10-27T07:09:31Z) - Contrastive General Graph Matching with Adaptive Augmentation Sampling [5.3459881796368505]
We introduce a novel Graph-centric Contrastive framework for Graph Matching (GCGM)
GCGM capitalizes on a vast pool of graph augmentations for contrastive learning, yet without needing any side information.
Our GCGM surpasses state-of-the-art self-supervised methods across various datasets.
arXiv Detail & Related papers (2024-06-25T01:08:03Z) - Through the Dual-Prism: A Spectral Perspective on Graph Data
Augmentation for Graph Classification [71.36575018271405]
We introduce the Dual-Prism (DP) augmentation method, comprising DP-Noise and DP-Mask.
We find that keeping the low-frequency eigenvalues unchanged can preserve the critical properties at a large scale when generating augmented graphs.
arXiv Detail & Related papers (2024-01-18T12:58:53Z) - Spectral Augmentations for Graph Contrastive Learning [50.149996923976836]
Contrastive learning has emerged as a premier method for learning representations with or without supervision.
Recent studies have shown its utility in graph representation learning for pre-training.
We propose a set of well-motivated graph transformation operations to provide a bank of candidates when constructing augmentations for a graph contrastive objective.
arXiv Detail & Related papers (2023-02-06T16:26:29Z) - Coarse-to-Fine Contrastive Learning on Graphs [38.41992365090377]
A variety of graph augmentation strategies have been employed to learn node representations in a self-supervised manner.
We introduce a self-ranking paradigm to ensure that the discriminative information among different nodes can be maintained.
Experiment results on various benchmark datasets verify the effectiveness of our algorithm.
arXiv Detail & Related papers (2022-12-13T08:17:20Z) - Graph Contrastive Learning with Implicit Augmentations [36.57536688367965]
Implicit Graph Contrastive Learning (iGCL) uses augmentations in latent space learned from a Variational Graph Auto-Encoder by reconstructing graph topological structure.
Experimental results on both graph-level and node-level tasks show that the proposed method achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-11-07T17:34:07Z) - ARIEL: Adversarial Graph Contrastive Learning [51.14695794459399]
ARIEL consistently outperforms the current graph contrastive learning methods for both node-level and graph-level classification tasks.
ARIEL is more robust in the face of adversarial attacks.
arXiv Detail & Related papers (2022-08-15T01:24:42Z) - Similarity-aware Positive Instance Sampling for Graph Contrastive
Pre-training [82.68805025636165]
We propose to select positive graph instances directly from existing graphs in the training set.
Our selection is based on certain domain-specific pair-wise similarity measurements.
Besides, we develop an adaptive node-level pre-training method to dynamically mask nodes to distribute them evenly in the graph.
arXiv Detail & Related papers (2022-06-23T20:12:51Z) - Adversarial Graph Contrastive Learning with Information Regularization [51.14695794459399]
Contrastive learning is an effective method in graph representation learning.
Data augmentation on graphs is far less intuitive and much harder to provide high-quality contrastive samples.
We propose a simple but effective method, Adversarial Graph Contrastive Learning (ARIEL)
It consistently outperforms the current graph contrastive learning methods in the node classification task over various real-world datasets.
arXiv Detail & Related papers (2022-02-14T05:54:48Z) - Graph Contrastive Learning with Augmentations [109.23158429991298]
We propose a graph contrastive learning (GraphCL) framework for learning unsupervised representations of graph data.
We show that our framework can produce graph representations of similar or better generalizability, transferrability, and robustness compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-10-22T20:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.