When Contrastive Learning Meets Active Learning: A Novel Graph Active
Learning Paradigm with Self-Supervision
- URL: http://arxiv.org/abs/2010.16091v2
- Date: Fri, 16 Apr 2021 12:28:32 GMT
- Title: When Contrastive Learning Meets Active Learning: A Novel Graph Active
Learning Paradigm with Self-Supervision
- Authors: Yanqiao Zhu and Weizhi Xu and Qiang Liu and Shu Wu
- Abstract summary: This paper studies active learning (AL) on graphs, whose purpose is to discover the most informative nodes to maximize the performance of graph neural networks (GNNs)
Motivated by the success of contrastive learning (CL), we propose a novel paradigm that seamlessly integrates graph AL with CL.
Comprehensive, confounding-free experiments on five public datasets demonstrate the superiority of our method over state-of-the-arts.
- Score: 19.938379604834743
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies active learning (AL) on graphs, whose purpose is to
discover the most informative nodes to maximize the performance of graph neural
networks (GNNs). Previously, most graph AL methods focus on learning node
representations from a carefully selected labeled dataset with large amount of
unlabeled data neglected. Motivated by the success of contrastive learning
(CL), we propose a novel paradigm that seamlessly integrates graph AL with CL.
While being able to leverage the power of abundant unlabeled data in a
self-supervised manner, nodes selected by AL further provide semantic
information that can better guide representation learning. Besides, previous
work measures the informativeness of nodes without considering the neighborhood
propagation scheme of GNNs, so that noisy nodes may be selected. We argue that
due to the smoothing nature of GNNs, the central nodes from homophilous
subgraphs should benefit the model training most. To this end, we present a
minimax selection scheme that explicitly harnesses neighborhood information and
discover homophilous subgraphs to facilitate active selection. Comprehensive,
confounding-free experiments on five public datasets demonstrate the
superiority of our method over state-of-the-arts.
Related papers
- Label Deconvolution for Node Representation Learning on Large-scale
Attributed Graphs against Learning Bias [75.44877675117749]
We propose an efficient label regularization technique, namely Label Deconvolution (LD), to alleviate the learning bias by a novel and highly scalable approximation to the inverse mapping of GNNs.
Experiments demonstrate LD significantly outperforms state-of-the-art methods on Open Graph datasets Benchmark.
arXiv Detail & Related papers (2023-09-26T13:09:43Z) - SMARTQUERY: An Active Learning Framework for Graph Neural Networks
through Hybrid Uncertainty Reduction [25.77052028238513]
We propose a framework to learn a graph neural network with very few labeled nodes using a hybrid uncertainty reduction function.
We demonstrate the competitive performance of our method against state-of-the-arts on very few labeled data.
arXiv Detail & Related papers (2022-12-02T20:49:38Z) - Learning with Few Labeled Nodes via Augmented Graph Self-Training [36.97506256446519]
A GST (Augmented Graph Self-Training) framework is built with two new (i.e., structural and semantic) augmentation modules on top of a decoupled GST backbone.
We investigate whether this novel framework can learn an effective graph predictive model with extremely limited labeled nodes.
arXiv Detail & Related papers (2022-08-26T03:36:01Z) - Node Representation Learning in Graph via Node-to-Neighbourhood Mutual
Information Maximization [27.701736055800314]
Key towards learning informative node representations in graphs lies in how to gain contextual information from the neighbourhood.
We present a self-supervised node representation learning strategy via directly maximizing the mutual information between the hidden representations of nodes and their neighbourhood.
Our framework is optimized via a surrogate contrastive loss, where the positive selection underpins the quality and efficiency of representation learning.
arXiv Detail & Related papers (2022-03-23T08:21:10Z) - Partition-Based Active Learning for Graph Neural Networks [17.386869902409153]
We study the problem of semi-supervised learning with Graph Neural Networks (GNNs) in an active learning setup.
We propose GraphPart, a novel partition-based active learning approach for GNNs.
arXiv Detail & Related papers (2022-01-23T22:51:14Z) - Uniting Heterogeneity, Inductiveness, and Efficiency for Graph
Representation Learning [68.97378785686723]
graph neural networks (GNNs) have greatly advanced the performance of node representation learning on graphs.
A majority class of GNNs are only designed for homogeneous graphs, leading to inferior adaptivity to the more informative heterogeneous graphs.
We propose a novel inductive, meta path-free message passing scheme that packs up heterogeneous node features with their associated edges from both low- and high-order neighbor nodes.
arXiv Detail & Related papers (2021-04-04T23:31:39Z) - Learning the Implicit Semantic Representation on Graph-Structured Data [57.670106959061634]
Existing representation learning methods in graph convolutional networks are mainly designed by describing the neighborhood of each node as a perceptual whole.
We propose a Semantic Graph Convolutional Networks (SGCN) that explores the implicit semantics by learning latent semantic-paths in graphs.
arXiv Detail & Related papers (2021-01-16T16:18:43Z) - Node2Seq: Towards Trainable Convolutions in Graph Neural Networks [59.378148590027735]
We propose a graph network layer, known as Node2Seq, to learn node embeddings with explicitly trainable weights for different neighboring nodes.
For a target node, our method sorts its neighboring nodes via attention mechanism and then employs 1D convolutional neural networks (CNNs) to enable explicit weights for information aggregation.
In addition, we propose to incorporate non-local information for feature learning in an adaptive manner based on the attention scores.
arXiv Detail & Related papers (2021-01-06T03:05:37Z) - Deep Reinforcement Learning of Graph Matching [63.469961545293756]
Graph matching (GM) under node and pairwise constraints has been a building block in areas from optimization to computer vision.
We present a reinforcement learning solver for GM i.e. RGM that seeks the node correspondence between pairwise graphs.
Our method differs from the previous deep graph matching model in the sense that they are focused on the front-end feature extraction and affinity function learning.
arXiv Detail & Related papers (2020-12-16T13:48:48Z) - Cyclic Label Propagation for Graph Semi-supervised Learning [52.102251202186025]
We introduce a novel framework for graph semi-supervised learning called CycProp.
CycProp integrates GNNs into the process of label propagation in a cyclic and mutually reinforcing manner.
In particular, our proposed CycProp updates the node embeddings learned by GNN module with the augmented information by label propagation.
arXiv Detail & Related papers (2020-11-24T02:55:40Z) - CAGNN: Cluster-Aware Graph Neural Networks for Unsupervised Graph
Representation Learning [19.432449825536423]
Unsupervised graph representation learning aims to learn low-dimensional node embeddings without supervision.
We present a novel cluster-aware graph neural network (CAGNN) model for unsupervised graph representation learning using self-supervised techniques.
arXiv Detail & Related papers (2020-09-03T13:57:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.