Integrating Transductive And Inductive Embeddings Improves Link
Prediction Accuracy
- URL: http://arxiv.org/abs/2108.10108v1
- Date: Mon, 23 Aug 2021 12:24:20 GMT
- Title: Integrating Transductive And Inductive Embeddings Improves Link
Prediction Accuracy
- Authors: Chitrank Gupta, Yash Jain, Abir De, Soumen Chakrabarti
- Abstract summary: In inductive graph embedding models, emphviz., graph neural networks (GNNs) have become increasingly accurate at link prediction (LP) in online social networks.
We demonstrate that, for a wide variety of GNN variants, node representation vectors obtained from Node2Vec serve as high quality input features to GNNs.
- Score: 24.306445780189005
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, inductive graph embedding models, \emph{viz.}, graph neural
networks (GNNs) have become increasingly accurate at link prediction (LP) in
online social networks. The performance of such networks depends strongly on
the input node features, which vary across networks and applications. Selecting
appropriate node features remains application-dependent and generally an open
question. Moreover, owing to privacy and ethical issues, use of personalized
node features is often restricted. In fact, many publicly available data from
online social network do not contain any node features (e.g., demography). In
this work, we provide a comprehensive experimental analysis which shows that
harnessing a transductive technique (e.g., Node2Vec) for obtaining initial node
representations, after which an inductive node embedding technique takes over,
leads to substantial improvements in link prediction accuracy. We demonstrate
that, for a wide variety of GNN variants, node representation vectors obtained
from Node2Vec serve as high quality input features to GNNs, thereby improving
LP performance.
Related papers
- Can GNNs Learn Link Heuristics? A Concise Review and Evaluation of Link Prediction Methods [16.428742189544955]
This paper explores the ability of Graph Neural Networks (GNNs) in learning various forms of information for link prediction.
Our analysis reveals that GNNs cannot effectively learn structural information related to the number of common neighbors between two nodes.
Also, our extensive experiments indicate that trainable node embeddings can improve the performance of GNN-based link prediction models.
arXiv Detail & Related papers (2024-11-22T03:38:20Z) - Degree-based stratification of nodes in Graph Neural Networks [66.17149106033126]
We modify the Graph Neural Network (GNN) architecture so that the weight matrices are learned, separately, for the nodes in each group.
This simple-to-implement modification seems to improve performance across datasets and GNN methods.
arXiv Detail & Related papers (2023-12-16T14:09:23Z) - Graph Neural Network Based Node Deployment for Throughput Enhancement [20.56966053013759]
We propose a novel graph neural network (GNN) method for the network node deployment problem.
We show that an expressive GNN has the capacity to approximate both the function value and the traffic permutation, as a theoretic support for the proposed method.
arXiv Detail & Related papers (2022-08-19T08:06:28Z) - Invertible Neural Networks for Graph Prediction [22.140275054568985]
In this work, we address conditional generation using deep invertible neural networks.
We adopt an end-to-end training approach since our objective is to address prediction and generation in the forward and backward processes at once.
arXiv Detail & Related papers (2022-06-02T17:28:33Z) - Identity-aware Graph Neural Networks [63.6952975763946]
We develop a class of message passing Graph Neural Networks (ID-GNNs) with greater expressive power than the 1-WL test.
ID-GNN extends existing GNN architectures by inductively considering nodes' identities during message passing.
We show that transforming existing GNNs to ID-GNNs yields on average 40% accuracy improvement on challenging node, edge, and graph property prediction tasks.
arXiv Detail & Related papers (2021-01-25T18:59:01Z) - Node2Seq: Towards Trainable Convolutions in Graph Neural Networks [59.378148590027735]
We propose a graph network layer, known as Node2Seq, to learn node embeddings with explicitly trainable weights for different neighboring nodes.
For a target node, our method sorts its neighboring nodes via attention mechanism and then employs 1D convolutional neural networks (CNNs) to enable explicit weights for information aggregation.
In addition, we propose to incorporate non-local information for feature learning in an adaptive manner based on the attention scores.
arXiv Detail & Related papers (2021-01-06T03:05:37Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Permutation-equivariant and Proximity-aware Graph Neural Networks with
Stochastic Message Passing [88.30867628592112]
Graph neural networks (GNNs) are emerging machine learning models on graphs.
Permutation-equivariance and proximity-awareness are two important properties highly desirable for GNNs.
We show that existing GNNs, mostly based on the message-passing mechanism, cannot simultaneously preserve the two properties.
In order to preserve node proximities, we augment the existing GNNs with node representations.
arXiv Detail & Related papers (2020-09-05T16:46:56Z) - Binarized Graph Neural Network [65.20589262811677]
We develop a binarized graph neural network to learn the binary representations of the nodes with binary network parameters.
Our proposed method can be seamlessly integrated into the existing GNN-based embedding approaches.
Experiments indicate that the proposed binarized graph neural network, namely BGN, is orders of magnitude more efficient in terms of both time and space.
arXiv Detail & Related papers (2020-04-19T09:43:14Z) - Generalization and Representational Limits of Graph Neural Networks [46.20253808402385]
We prove that several important graph properties cannot be computed by graph neural networks (GNNs) that rely entirely on local information.
We provide the first data dependent generalization bounds for message passing GNNs.
Our bounds are much tighter than existing VC-dimension based guarantees for GNNs, and are comparable to Rademacher bounds for recurrent neural networks.
arXiv Detail & Related papers (2020-02-14T18:10:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.