Jointly Learnable Data Augmentations for Self-Supervised GNNs
- URL: http://arxiv.org/abs/2108.10420v1
- Date: Mon, 23 Aug 2021 21:33:12 GMT
- Title: Jointly Learnable Data Augmentations for Self-Supervised GNNs
- Authors: Zekarias T. Kefato and Sarunas Girdzijauskas and Hannes St\"ark
- Abstract summary: We propose GraphSurgeon, a novel self-supervised learning method for graph representation learning.
We take advantage of the flexibility of the learnable data augmentation and introduce a new strategy that augments in the embedding space.
Our finding shows that GraphSurgeon is comparable to six SOTA semi-supervised and on par with five SOTA self-supervised baselines in node classification tasks.
- Score: 0.311537581064266
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Self-supervised Learning (SSL) aims at learning representations of objects
without relying on manual labeling. Recently, a number of SSL methods for graph
representation learning have achieved performance comparable to SOTA
semi-supervised GNNs. A Siamese network, which relies on data augmentation, is
the popular architecture used in these methods. However, these methods rely on
heuristically crafted data augmentation techniques. Furthermore, they use
either contrastive terms or other tricks (e.g., asymmetry) to avoid trivial
solutions that can occur in Siamese networks. In this study, we propose,
GraphSurgeon, a novel SSL method for GNNs with the following features. First,
instead of heuristics we propose a learnable data augmentation method that is
jointly learned with the embeddings by leveraging the inherent signal encoded
in the graph. In addition, we take advantage of the flexibility of the
learnable data augmentation and introduce a new strategy that augments in the
embedding space, called post augmentation. This strategy has a significantly
lower memory overhead and run-time cost. Second, as it is difficult to sample
truly contrastive terms, we avoid explicit negative sampling. Third, instead of
relying on engineering tricks, we use a scalable constrained optimization
objective motivated by Laplacian Eigenmaps to avoid trivial solutions. To
validate the practical use of GraphSurgeon, we perform empirical evaluation
using 14 public datasets across a number of domains and ranging from small to
large scale graphs with hundreds of millions of edges. Our finding shows that
GraphSurgeon is comparable to six SOTA semi-supervised and on par with five
SOTA self-supervised baselines in node classification tasks. The source code is
available at https://github.com/zekarias-tilahun/graph-surgeon.
Related papers
- Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - GRAPES: Learning to Sample Graphs for Scalable Graph Neural Networks [2.4175455407547015]
Graph neural networks learn to represent nodes by aggregating information from their neighbors.
Several existing methods address this by sampling a small subset of nodes, scaling GNNs to much larger graphs.
We introduce GRAPES, an adaptive sampling method that learns to identify the set of nodes crucial for training a GNN.
arXiv Detail & Related papers (2023-10-05T09:08:47Z) - SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - Training Graph Neural Networks on Growing Stochastic Graphs [114.75710379125412]
Graph Neural Networks (GNNs) rely on graph convolutions to exploit meaningful patterns in networked data.
We propose to learn GNNs on very large graphs by leveraging the limit object of a sequence of growing graphs, the graphon.
arXiv Detail & Related papers (2022-10-27T16:00:45Z) - SizeShiftReg: a Regularization Method for Improving Size-Generalization
in Graph Neural Networks [5.008597638379227]
Graph neural networks (GNNs) have become the de facto model of choice for graph classification.
We propose a regularization strategy that can be applied to any GNN to improve its generalization capabilities without requiring access to the test data.
Our regularization is based on the idea of simulating a shift in the size of the training graphs using coarsening techniques.
arXiv Detail & Related papers (2022-07-16T09:50:45Z) - Graph Contrastive Learning with Augmentations [109.23158429991298]
We propose a graph contrastive learning (GraphCL) framework for learning unsupervised representations of graph data.
We show that our framework can produce graph representations of similar or better generalizability, transferrability, and robustness compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-10-22T20:13:43Z) - Robust Optimization as Data Augmentation for Large-scale Graphs [117.2376815614148]
We propose FLAG (Free Large-scale Adversarial Augmentation on Graphs), which iteratively augments node features with gradient-based adversarial perturbations during training.
FLAG is a general-purpose approach for graph data, which universally works in node classification, link prediction, and graph classification tasks.
arXiv Detail & Related papers (2020-10-19T21:51:47Z) - Self-supervised Learning on Graphs: Deep Insights and New Direction [66.78374374440467]
Self-supervised learning (SSL) aims to create domain specific pretext tasks on unlabeled data.
There are increasing interests in generalizing deep learning to the graph domain in the form of graph neural networks (GNNs)
arXiv Detail & Related papers (2020-06-17T20:30:04Z) - Isometric Graph Neural Networks [5.306334746787569]
We propose a technique to learn Isometric Graph Neural Networks (IGNN)
IGNN requires changing the input representation space and loss function to enable any GNN algorithm to generate representations that reflect distances between nodes.
We observe a consistent and substantial improvement as high as 400% in Kendall's Tau (KT)
arXiv Detail & Related papers (2020-06-16T22:51:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.