Augmentations in Graph Contrastive Learning: Current Methodological
Flaws & Towards Better Practices
- URL: http://arxiv.org/abs/2111.03220v1
- Date: Fri, 5 Nov 2021 02:15:01 GMT
- Title: Augmentations in Graph Contrastive Learning: Current Methodological
Flaws & Towards Better Practices
- Authors: Puja Trivedi, Ekdeep Singh Lubana, Yujun Yan, Yaoqing Yang, Danai
Koutra
- Abstract summary: Graph classification has applications in bioinformatics, social sciences, automated fake news detection, web document classification, and more.
Recently, contrastive learning (CL) has enabled unsupervised computer vision models to compete well against supervised ones.
Motivated by these discrepancies, we seek to determine: (i) why existing graph CL frameworks perform well despite weak augmentations and limited data; and (ii) whether adhering to visual CL principles can improve performance on graph classification tasks.
- Score: 20.95255742208036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph classification has applications in bioinformatics, social sciences,
automated fake news detection, web document classification, and more. In many
practical scenarios, including web-scale applications, where labels are scarce
or hard to obtain, unsupervised learning is a natural paradigm but it trades
off performance. Recently, contrastive learning (CL) has enabled unsupervised
computer vision models to compete well against supervised ones. Theoretical and
empirical works analyzing visual CL frameworks find that leveraging large
datasets and domain aware augmentations is essential for framework success.
Interestingly, graph CL frameworks often report high performance while using
orders of magnitude smaller data, and employing domain-agnostic augmentations
(e.g., node or edge dropping, feature perturbations) that can corrupt the
graphs' underlying properties.
Motivated by these discrepancies, we seek to determine: (i) why existing
graph CL frameworks perform well despite weak augmentations and limited data;
and (ii) whether adhering to visual CL principles can improve performance on
graph classification tasks. Through extensive analysis, we identify flawed
practices in graph data augmentation and evaluation protocols that are commonly
used in the graph CL literature, and propose improved practices and sanity
checks for future research and applications. We show that on small benchmark
datasets, the inductive bias of graph neural networks can significantly
compensate for the limitations of existing frameworks. In case studies with
relatively larger graph classification tasks, we find that commonly used
domain-agnostic augmentations perform poorly, while adhering to principles in
visual CL can significantly improve performance. For example, in graph-based
document classification, which can be used for better web search, we show
task-relevant augmentations improve accuracy by 20%.
Related papers
- Let Invariant Rationale Discovery Inspire Graph Contrastive Learning [98.10268114789775]
We argue that a high-performing augmentation should preserve the salient semantics of anchor graphs regarding instance-discrimination.
We propose a new framework, Rationale-aware Graph Contrastive Learning (RGCL)
RGCL uses a rationale generator to reveal salient features about graph instance-discrimination as the rationale, and then creates rationale-aware views for contrastive learning.
arXiv Detail & Related papers (2022-06-16T01:28:40Z) - Towards Unsupervised Deep Graph Structure Learning [67.58720734177325]
We propose an unsupervised graph structure learning paradigm, where the learned graph topology is optimized by data itself without any external guidance.
Specifically, we generate a learning target from the original data as an "anchor graph", and use a contrastive loss to maximize the agreement between the anchor graph and the learned graph.
arXiv Detail & Related papers (2022-01-17T11:57:29Z) - Graph Contrastive Learning Automated [94.41860307845812]
Graph contrastive learning (GraphCL) has emerged with promising representation learning performance.
The effectiveness of GraphCL hinges on ad-hoc data augmentations, which have to be manually picked per dataset.
This paper proposes a unified bi-level optimization framework to automatically, adaptively and dynamically select data augmentations when performing GraphCL on specific graph data.
arXiv Detail & Related papers (2021-06-10T16:35:27Z) - Structure-Enhanced Meta-Learning For Few-Shot Graph Classification [53.54066611743269]
This work explores the potential of metric-based meta-learning for solving few-shot graph classification.
An implementation upon GIN, named SMFGIN, is tested on two datasets, Chembl and TRIANGLES.
arXiv Detail & Related papers (2021-03-05T09:03:03Z) - Model-Agnostic Graph Regularization for Few-Shot Learning [60.64531995451357]
We present a comprehensive study on graph embedded few-shot learning.
We introduce a graph regularization approach that allows a deeper understanding of the impact of incorporating graph information between labels.
Our approach improves the performance of strong base learners by up to 2% on Mini-ImageNet and 6.7% on ImageNet-FS.
arXiv Detail & Related papers (2021-02-14T05:28:13Z) - Graph Contrastive Learning with Adaptive Augmentation [23.37786673825192]
We propose a novel graph contrastive representation learning method with adaptive augmentation.
Specifically, we design augmentation schemes based on node centrality measures to highlight important connective structures.
Our proposed method consistently outperforms existing state-of-the-art baselines and even surpasses some supervised counterparts.
arXiv Detail & Related papers (2020-10-27T15:12:21Z) - Active Learning on Attributed Graphs via Graph Cognizant Logistic
Regression and Preemptive Query Generation [37.742218733235084]
We propose a novel graph-based active learning algorithm for the task of node classification in attributed graphs.
Our algorithm uses graph cognizant logistic regression, equivalent to a linearized graph convolutional neural network (GCN) for the prediction phase and maximizes the expected error reduction in the query phase.
We conduct experiments on five public benchmark datasets, demonstrating a significant improvement over state-of-the-art approaches.
arXiv Detail & Related papers (2020-07-09T18:00:53Z) - Quantifying Challenges in the Application of Graph Representation
Learning [0.0]
We provide an application oriented perspective to a set of popular embedding approaches.
We evaluate their representational power with respect to real-world graph properties.
Our results suggest that "one-to-fit-all" GRL approaches are hard to define in real-world scenarios.
arXiv Detail & Related papers (2020-06-18T03:19:43Z) - GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training [62.73470368851127]
Graph representation learning has emerged as a powerful technique for addressing real-world problems.
We design Graph Contrastive Coding -- a self-supervised graph neural network pre-training framework.
We conduct experiments on three graph learning tasks and ten graph datasets.
arXiv Detail & Related papers (2020-06-17T16:18:35Z) - Data Augmentation for Graph Neural Networks [32.24311481878144]
We study graph data augmentation for graph neural networks (GNNs) in the context of improving semi-supervised node-classification.
Our work shows that neural edge predictors can effectively encode class-homophilic structure to promote intra-class edges and demote inter-class edges in given graph structure.
Our main contribution introduces the GAug graph data augmentation framework, which leverages these insights to improve performance in GNN-based node classification via edge prediction.
arXiv Detail & Related papers (2020-06-11T21:17:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.