Improving Graph Neural Networks at Scale: Combining Approximate PageRank
and CoreRank
- URL: http://arxiv.org/abs/2211.04248v1
- Date: Tue, 8 Nov 2022 13:51:49 GMT
- Title: Improving Graph Neural Networks at Scale: Combining Approximate PageRank
and CoreRank
- Authors: Ariel R. Ramos Vela, Johannes F. Lutzeyer, Anastasios Giovanidis,
Michalis Vazirgiannis
- Abstract summary: We propose a scalable solution to propagate information on Graph Neural Networks (GNNs)
The CorePPR model uses a learnable convex combination of the approximate personalised PageRank and the CoreRank to diffuse multi-hop neighbourhood information in GNNs.
We demonstrate that CorePPR outperforms PPRGo on large graphs where selecting the most influential nodes is particularly relevant for scalability.
- Score: 20.948992161528466
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Graph Neural Networks (GNNs) have achieved great successes in many learning
tasks performed on graph structures. Nonetheless, to propagate information GNNs
rely on a message passing scheme which can become prohibitively expensive when
working with industrial-scale graphs. Inspired by the PPRGo model, we propose
the CorePPR model, a scalable solution that utilises a learnable convex
combination of the approximate personalised PageRank and the CoreRank to
diffuse multi-hop neighbourhood information in GNNs. Additionally, we
incorporate a dynamic mechanism to select the most influential neighbours for a
particular node which reduces training time while preserving the performance of
the model. Overall, we demonstrate that CorePPR outperforms PPRGo, particularly
on large graphs where selecting the most influential nodes is particularly
relevant for scalability. Our code is publicly available at:
https://github.com/arielramos97/CorePPR.
Related papers
- Spectral Greedy Coresets for Graph Neural Networks [61.24300262316091]
The ubiquity of large-scale graphs in node-classification tasks hinders the real-world applications of Graph Neural Networks (GNNs)
This paper studies graph coresets for GNNs and avoids the interdependence issue by selecting ego-graphs based on their spectral embeddings.
Our spectral greedy graph coreset (SGGC) scales to graphs with millions of nodes, obviates the need for model pre-training, and applies to low-homophily graphs.
arXiv Detail & Related papers (2024-05-27T17:52:12Z) - SPGNN: Recognizing Salient Subgraph Patterns via Enhanced Graph Convolution and Pooling [25.555741218526464]
Graph neural networks (GNNs) have revolutionized the field of machine learning on non-Euclidean data such as graphs and networks.
We propose a concatenation-based graph convolution mechanism that injectively updates node representations.
We also design a novel graph pooling module, called WL-SortPool, to learn important subgraph patterns in a deep-learning manner.
arXiv Detail & Related papers (2024-04-21T13:11:59Z) - Graph Ladling: Shockingly Simple Parallel GNN Training without
Intermediate Communication [100.51884192970499]
GNNs are a powerful family of neural networks for learning over graphs.
scaling GNNs either by deepening or widening suffers from prevalent issues of unhealthy gradients, over-smoothening, information squashing.
We propose not to deepen or widen current GNNs, but instead present a data-centric perspective of model soups tailored for GNNs.
arXiv Detail & Related papers (2023-06-18T03:33:46Z) - Dynamic Graph Message Passing Networks for Visual Recognition [112.49513303433606]
Modelling long-range dependencies is critical for scene understanding tasks in computer vision.
A fully-connected graph is beneficial for such modelling, but its computational overhead is prohibitive.
We propose a dynamic graph message passing network, that significantly reduces the computational complexity.
arXiv Detail & Related papers (2022-09-20T14:41:37Z) - HPGNN: Using Hierarchical Graph Neural Networks for Outdoor Point Cloud
Processing [0.7649716717097428]
Motivated by recent improvements in point cloud processing for autonomous navigation, we focus on using hierarchical graph neural networks for processing.
We propose Hierarchical Point Graph Neural Network (HPGNN)
It learns node features at various levels of graph coarseness to extract information.
This enables to learn over a large point cloud while retaining fine details that existing point-level graph networks struggle to achieve.
arXiv Detail & Related papers (2022-06-05T11:18:09Z) - Personalized PageRank Graph Attention Networks [0.0]
A graph neural network (GNN) is a framework to learn from graph-structured data.
GNNs typically only use the information of a very limited neighborhood for each node to avoid over-smoothing.
In this work, we incorporate the limit distribution of Personalized PageRank (PPR) into graph attention networks (GATs) to reflect the larger neighbor information without introducing over-smoothing.
arXiv Detail & Related papers (2022-05-27T22:36:47Z) - Inferential SIR-GN: Scalable Graph Representation Learning [0.4699313647907615]
Graph representation learning methods generate numerical vector representations for the nodes in a network.
In this work, we propose Inferential SIR-GN, a model which is pre-trained on random graphs, then computes node representations rapidly.
We demonstrate that the model is able to capture node's structural role information, and show excellent performance at node and graph classification tasks, on unseen networks.
arXiv Detail & Related papers (2021-11-08T20:56:37Z) - Distance Encoding: Design Provably More Powerful Neural Networks for
Graph Representation Learning [63.97983530843762]
Graph Neural Networks (GNNs) have achieved great success in graph representation learning.
GNNs generate identical representations for graph substructures that may in fact be very different.
More powerful GNNs, proposed recently by mimicking higher-order tests, are inefficient as they cannot sparsity of underlying graph structure.
We propose Distance Depiction (DE) as a new class of graph representation learning.
arXiv Detail & Related papers (2020-08-31T23:15:40Z) - Scaling Graph Neural Networks with Approximate PageRank [64.92311737049054]
We present the PPRGo model which utilizes an efficient approximation of information diffusion in GNNs.
In addition to being faster, PPRGo is inherently scalable, and can be trivially parallelized for large datasets like those found in industry settings.
We show that training PPRGo and predicting labels for all nodes in this graph takes under 2 minutes on a single machine, far outpacing other baselines on the same graph.
arXiv Detail & Related papers (2020-07-03T09:30:07Z) - Policy-GNN: Aggregation Optimization for Graph Neural Networks [60.50932472042379]
Graph neural networks (GNNs) aim to model the local graph structures and capture the hierarchical patterns by aggregating the information from neighbors.
It is a challenging task to develop an effective aggregation strategy for each node, given complex graphs and sparse features.
We propose Policy-GNN, a meta-policy framework that models the sampling procedure and message passing of GNNs into a combined learning process.
arXiv Detail & Related papers (2020-06-26T17:03:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.