Using Graph Algorithms to Pretrain Graph Completion Transformers
- URL: http://arxiv.org/abs/2210.07453v2
- Date: Mon, 27 Mar 2023 15:04:30 GMT
- Title: Using Graph Algorithms to Pretrain Graph Completion Transformers
- Authors: Jonathan Pilault, Michael Galkin, Bahare Fatemi, Perouz Taslakian,
David Vasquez, Christopher Pal
- Abstract summary: Self-supervised pretraining can enhance performance on downstream graph, link, and node classification tasks.
We investigate five different pretraining signals, constructed using several graph algorithms and no external data, as well as their combination.
We propose a new path-finding algorithm guided by information gain and find that it is the best-performing pretraining task.
- Score: 8.327657957422833
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent work on Graph Neural Networks has demonstrated that self-supervised
pretraining can further enhance performance on downstream graph, link, and node
classification tasks. However, the efficacy of pretraining tasks has not been
fully investigated for downstream large knowledge graph completion tasks. Using
a contextualized knowledge graph embedding approach, we investigate five
different pretraining signals, constructed using several graph algorithms and
no external data, as well as their combination. We leverage the versatility of
our Transformer-based model to explore graph structure generation pretraining
tasks (i.e. path and k-hop neighborhood generation), typically inapplicable to
most graph embedding methods. We further propose a new path-finding algorithm
guided by information gain and find that it is the best-performing pretraining
task across three downstream knowledge graph completion datasets. While using
our new path-finding algorithm as a pretraining signal provides 2-3% MRR
improvements, we show that pretraining on all signals together gives the best
knowledge graph completion results. In a multitask setting that combines all
pretraining tasks, our method surpasses the latest and strong performing
knowledge graph embedding methods on all metrics for FB15K-237, on MRR and
Hit@1 for WN18RRand on MRR and hit@10 for JF17K (a knowledge hypergraph
dataset).
Related papers
- SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - All in One: Multi-task Prompting for Graph Neural Networks [30.457491401821652]
We propose a novel multi-task prompting method for graph models.
We first unify the format of graph prompts and language prompts with the prompt token, token structure, and inserting pattern.
We then study the task space of various graph applications and reformulate downstream problems to the graph-level task.
arXiv Detail & Related papers (2023-07-04T06:27:31Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - A Robust and Generalized Framework for Adversarial Graph Embedding [73.37228022428663]
We propose a robust framework for adversarial graph embedding, named AGE.
AGE generates the fake neighbor nodes as the enhanced negative samples from the implicit distribution.
Based on this framework, we propose three models to handle three types of graph data.
arXiv Detail & Related papers (2021-05-22T07:05:48Z) - Co-embedding of Nodes and Edges with Graph Neural Networks [13.020745622327894]
Graph embedding is a way to transform and encode the data structure in high dimensional and non-Euclidean feature space.
CensNet is a general graph embedding framework, which embeds both nodes and edges to a latent feature space.
Our approach achieves or matches the state-of-the-art performance in four graph learning tasks.
arXiv Detail & Related papers (2020-10-25T22:39:31Z) - Graph Contrastive Learning with Augmentations [109.23158429991298]
We propose a graph contrastive learning (GraphCL) framework for learning unsupervised representations of graph data.
We show that our framework can produce graph representations of similar or better generalizability, transferrability, and robustness compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-10-22T20:13:43Z) - GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training [62.73470368851127]
Graph representation learning has emerged as a powerful technique for addressing real-world problems.
We design Graph Contrastive Coding -- a self-supervised graph neural network pre-training framework.
We conduct experiments on three graph learning tasks and ten graph datasets.
arXiv Detail & Related papers (2020-06-17T16:18:35Z) - Heuristic Semi-Supervised Learning for Graph Generation Inspired by
Electoral College [80.67842220664231]
We propose a novel pre-processing technique, namely ELectoral COllege (ELCO), which automatically expands new nodes and edges to refine the label similarity within a dense subgraph.
In all setups tested, our method boosts the average score of base models by a large margin of 4.7 points, as well as consistently outperforms the state-of-the-art.
arXiv Detail & Related papers (2020-06-10T14:48:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.