ARIEL: Adversarial Graph Contrastive Learning
- URL: http://arxiv.org/abs/2208.06956v2
- Date: Tue, 6 Feb 2024 02:45:28 GMT
- Title: ARIEL: Adversarial Graph Contrastive Learning
- Authors: Shengyu Feng, Baoyu Jing, Yada Zhu, Hanghang Tong
- Abstract summary: ARIEL consistently outperforms the current graph contrastive learning methods for both node-level and graph-level classification tasks.
ARIEL is more robust in the face of adversarial attacks.
- Score: 51.14695794459399
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive learning is an effective unsupervised method in graph
representation learning, and the key component of contrastive learning lies in
the construction of positive and negative samples. Previous methods usually
utilize the proximity of nodes in the graph as the principle. Recently, the
data-augmentation-based contrastive learning method has advanced to show great
power in the visual domain, and some works extended this method from images to
graphs. However, unlike the data augmentation on images, the data augmentation
on graphs is far less intuitive and much harder to provide high-quality
contrastive samples, which leaves much space for improvement. In this work, by
introducing an adversarial graph view for data augmentation, we propose a
simple but effective method, Adversarial Graph Contrastive Learning (ARIEL), to
extract informative contrastive samples within reasonable constraints. We
develop a new technique called information regularization for stable training
and use subgraph sampling for scalability. We generalize our method from
node-level contrastive learning to the graph level by treating each graph
instance as a super-node. ARIEL consistently outperforms the current graph
contrastive learning methods for both node-level and graph-level classification
tasks on real-world datasets. We further demonstrate that ARIEL is more robust
in the face of adversarial attacks.
Related papers
- Uncovering Capabilities of Model Pruning in Graph Contrastive Learning [0.0]
We reformulate the problem of graph contrastive learning via contrasting different model versions rather than augmented views.
We extensively validate our method on various benchmarks regarding graph classification via unsupervised and transfer learning.
arXiv Detail & Related papers (2024-10-27T07:09:31Z) - Spectral Augmentations for Graph Contrastive Learning [50.149996923976836]
Contrastive learning has emerged as a premier method for learning representations with or without supervision.
Recent studies have shown its utility in graph representation learning for pre-training.
We propose a set of well-motivated graph transformation operations to provide a bank of candidates when constructing augmentations for a graph contrastive objective.
arXiv Detail & Related papers (2023-02-06T16:26:29Z) - Localized Contrastive Learning on Graphs [110.54606263711385]
We introduce a simple yet effective contrastive model named Localized Graph Contrastive Learning (Local-GCL)
In spite of its simplicity, Local-GCL achieves quite competitive performance in self-supervised node representation learning tasks on graphs with various scales and properties.
arXiv Detail & Related papers (2022-12-08T23:36:00Z) - Adversarial Graph Contrastive Learning with Information Regularization [51.14695794459399]
Contrastive learning is an effective method in graph representation learning.
Data augmentation on graphs is far less intuitive and much harder to provide high-quality contrastive samples.
We propose a simple but effective method, Adversarial Graph Contrastive Learning (ARIEL)
It consistently outperforms the current graph contrastive learning methods in the node classification task over various real-world datasets.
arXiv Detail & Related papers (2022-02-14T05:54:48Z) - Dual Space Graph Contrastive Learning [82.81372024482202]
We propose a novel graph contrastive learning method, namely textbfDual textbfSpace textbfGraph textbfContrastive (DSGC) Learning.
Since both spaces have their own advantages to represent graph data in the embedding spaces, we hope to utilize graph contrastive learning to bridge the spaces and leverage advantages from both sides.
arXiv Detail & Related papers (2022-01-19T04:10:29Z) - Bootstrapping Informative Graph Augmentation via A Meta Learning
Approach [21.814940639910358]
In graph contrastive learning, benchmark methods apply various graph augmentation approaches.
Most of the augmentation methods are non-learnable, which causes the issue of generating unbeneficial augmented graphs.
We motivate our method to generate augmented graph by a learnable graph augmenter, called MEta Graph Augmentation (MEGA)
arXiv Detail & Related papers (2022-01-11T07:15:13Z) - Multi-view Contrastive Graph Clustering [12.463334005083379]
We propose a generic framework to cluster multi-view attributed graph data.
Inspired by the success of contrastive learning, we propose multi-view contrastive graph clustering (MCGC) method.
Our simple approach outperforms existing deep learning-based methods.
arXiv Detail & Related papers (2021-10-22T15:22:42Z) - Edge but not Least: Cross-View Graph Pooling [76.71497833616024]
This paper presents a cross-view graph pooling (Co-Pooling) method to better exploit crucial graph structure information.
Through cross-view interaction, edge-view pooling and node-view pooling seamlessly reinforce each other to learn more informative graph-level representations.
arXiv Detail & Related papers (2021-09-24T08:01:23Z) - Sub-graph Contrast for Scalable Self-Supervised Graph Representation
Learning [21.0019144298605]
Existing graph neural networks fed with the complete graph data are not scalable due to limited computation and memory costs.
textscSubg-Con is proposed by utilizing the strong correlation between central nodes and their sampled subgraphs to capture regional structure information.
Compared with existing graph representation learning approaches, textscSubg-Con has prominent performance advantages in weaker supervision requirements, model learning scalability, and parallelization.
arXiv Detail & Related papers (2020-09-22T01:58:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.