Self-supervised Learning on Graphs: Deep Insights and New Direction
- URL: http://arxiv.org/abs/2006.10141v1
- Date: Wed, 17 Jun 2020 20:30:04 GMT
- Title: Self-supervised Learning on Graphs: Deep Insights and New Direction
- Authors: Wei Jin, Tyler Derr, Haochen Liu, Yiqi Wang, Suhang Wang, Zitao Liu,
Jiliang Tang
- Abstract summary: Self-supervised learning (SSL) aims to create domain specific pretext tasks on unlabeled data.
There are increasing interests in generalizing deep learning to the graph domain in the form of graph neural networks (GNNs)
- Score: 66.78374374440467
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The success of deep learning notoriously requires larger amounts of costly
annotated data. This has led to the development of self-supervised learning
(SSL) that aims to alleviate this limitation by creating domain specific
pretext tasks on unlabeled data. Simultaneously, there are increasing interests
in generalizing deep learning to the graph domain in the form of graph neural
networks (GNNs). GNNs can naturally utilize unlabeled nodes through the simple
neighborhood aggregation that is unable to thoroughly make use of unlabeled
nodes. Thus, we seek to harness SSL for GNNs to fully exploit the unlabeled
data. Different from data instances in the image and text domains, nodes in
graphs present unique structure information and they are inherently linked
indicating not independent and identically distributed (or i.i.d.). Such
complexity is a double-edged sword for SSL on graphs. On the one hand, it
determines that it is challenging to adopt solutions from the image and text
domains to graphs and dedicated efforts are desired. On the other hand, it
provides rich information that enables us to build SSL from a variety of
perspectives. Thus, in this paper, we first deepen our understandings on when,
why, and which strategies of SSL work with GNNs by empirically studying
numerous basic SSL pretext tasks on graphs. Inspired by deep insights from the
empirical studies, we propose a new direction SelfTask to build advanced
pretext tasks that are able to achieve state-of-the-art performance on various
real-world datasets. The specific experimental settings to reproduce our
results can be found in \url{https://github.com/ChandlerBang/SelfTask-GNN}.
Related papers
- Loss-aware Curriculum Learning for Heterogeneous Graph Neural Networks [30.333265803394998]
This paper investigates the application of curriculum learning techniques to improve the performance of Heterogeneous Graph Neural Networks (GNNs)
To better classify the quality of the data, we design a loss-aware training schedule, named LTS, that measures the quality of every nodes of the data.
Our findings demonstrate the efficacy of curriculum learning in enhancing HGNNs capabilities for analyzing complex graph-structured data.
arXiv Detail & Related papers (2024-02-29T05:44:41Z) - Every Node is Different: Dynamically Fusing Self-Supervised Tasks for
Attributed Graph Clustering [59.45743537594695]
We propose Dynamically Fusing Self-Supervised Learning (DyFSS) for graph clustering.
DyFSS fuses features extracted from diverse SSL tasks using distinct weights derived from a gating network.
Experiments show DyFSS outperforms state-of-the-art multi-task SSL methods by up to 8.66% on the accuracy metric.
arXiv Detail & Related papers (2024-01-12T14:24:10Z) - OpenGSL: A Comprehensive Benchmark for Graph Structure Learning [40.50100033304329]
We introduce OpenGSL, the first comprehensive benchmark for Graph Structure Learning (GSL)
OpenGSL enables a fair comparison among state-of-the-art GSL methods by evaluating them across various popular datasets.
We find that there is no significant correlation between the homophily of the learned structure and task performance, challenging the common belief.
arXiv Detail & Related papers (2023-06-17T07:22:25Z) - Learning with Few Labeled Nodes via Augmented Graph Self-Training [36.97506256446519]
A GST (Augmented Graph Self-Training) framework is built with two new (i.e., structural and semantic) augmentation modules on top of a decoupled GST backbone.
We investigate whether this novel framework can learn an effective graph predictive model with extremely limited labeled nodes.
arXiv Detail & Related papers (2022-08-26T03:36:01Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - Towards Unsupervised Deep Graph Structure Learning [67.58720734177325]
We propose an unsupervised graph structure learning paradigm, where the learned graph topology is optimized by data itself without any external guidance.
Specifically, we generate a learning target from the original data as an "anchor graph", and use a contrastive loss to maximize the agreement between the anchor graph and the learned graph.
arXiv Detail & Related papers (2022-01-17T11:57:29Z) - Jointly Learnable Data Augmentations for Self-Supervised GNNs [0.311537581064266]
We propose GraphSurgeon, a novel self-supervised learning method for graph representation learning.
We take advantage of the flexibility of the learnable data augmentation and introduce a new strategy that augments in the embedding space.
Our finding shows that GraphSurgeon is comparable to six SOTA semi-supervised and on par with five SOTA self-supervised baselines in node classification tasks.
arXiv Detail & Related papers (2021-08-23T21:33:12Z) - Graph Self-Supervised Learning: A Survey [73.86209411547183]
Self-supervised learning (SSL) has become a promising and trending learning paradigm for graph data.
We present a timely and comprehensive review of the existing approaches which employ SSL techniques for graph data.
arXiv Detail & Related papers (2021-02-27T03:04:21Z) - Contrastive and Generative Graph Convolutional Networks for Graph-based
Semi-Supervised Learning [64.98816284854067]
Graph-based Semi-Supervised Learning (SSL) aims to transfer the labels of a handful of labeled data to the remaining massive unlabeled data via a graph.
A novel GCN-based SSL algorithm is presented in this paper to enrich the supervision signals by utilizing both data similarities and graph structure.
arXiv Detail & Related papers (2020-09-15T13:59:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.