Large Scale Learning on Non-Homophilous Graphs: New Benchmarks and
Strong Simple Methods
- URL: http://arxiv.org/abs/2110.14446v1
- Date: Wed, 27 Oct 2021 14:02:41 GMT
- Title: Large Scale Learning on Non-Homophilous Graphs: New Benchmarks and
Strong Simple Methods
- Authors: Derek Lim, Felix Hohne, Xiuyu Li, Sijia Linda Huang, Vaishnavi Gupta,
Omkar Bhalerao, Ser-Nam Lim
- Abstract summary: New Graph Neural Networks (GNNs) have been developed that move beyond the homophily regime.
We introduce diverse non-homophilous datasets from a variety of application areas.
We show that existing scalable graph learning and graph minibatching techniques lead to performance degradation.
- Score: 16.170826632437183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many widely used datasets for graph machine learning tasks have generally
been homophilous, where nodes with similar labels connect to each other.
Recently, new Graph Neural Networks (GNNs) have been developed that move beyond
the homophily regime; however, their evaluation has often been conducted on
small graphs with limited application domains. We collect and introduce diverse
non-homophilous datasets from a variety of application areas that have up to
384x more nodes and 1398x more edges than prior datasets. We further show that
existing scalable graph learning and graph minibatching techniques lead to
performance degradation on these non-homophilous datasets, thus highlighting
the need for further work on scalable non-homophilous methods. To address these
concerns, we introduce LINKX -- a strong simple method that admits
straightforward minibatch training and inference. Extensive experimental
results with representative simple methods and GNNs across our proposed
datasets show that LINKX achieves state-of-the-art performance for learning on
non-homophilous graphs. Our codes and data are available at
https://github.com/CUAI/Non-Homophily-Large-Scale.
Related papers
- Non-Homophilic Graph Pre-Training and Prompt Learning [11.996173149569627]
We propose ProNoG, a novel pre-training and prompt learning framework for non-homophilic graphs.
First, we analyze existing graph pre-training methods, providing theoretical insights into the choice of pre-training tasks.
Second, recognizing that each node exhibits unique non-homophilic characteristics, we propose a conditional network to characterize the node-specific patterns in downstream tasks.
arXiv Detail & Related papers (2024-08-22T17:57:31Z) - The Heterophilic Graph Learning Handbook: Benchmarks, Models, Theoretical Analysis, Applications and Challenges [101.83124435649358]
Homophily principle, ie nodes with the same labels or similar attributes are more likely to be connected.
Recent work has identified a non-trivial set of datasets where GNN's performance compared to the NN's is not satisfactory.
arXiv Detail & Related papers (2024-07-12T18:04:32Z) - Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - A Simple and Scalable Graph Neural Network for Large Directed Graphs [11.792826520370774]
We investigate various combinations of node representations and edge direction awareness within an input graph.
In response, we propose a simple yet holistic classification method A2DUG.
We demonstrate that A2DUG stably performs well on various datasets and improves the accuracy up to 11.29 compared with the state-of-the-art methods.
arXiv Detail & Related papers (2023-06-14T06:24:58Z) - GCNH: A Simple Method For Representation Learning On Heterophilous
Graphs [4.051099980410583]
Graph Neural Networks (GNNs) are well-suited for learning on homophilous graphs.
Recent works have proposed extensions to standard GNN architectures to improve performance on heterophilous graphs.
We propose GCN for Heterophily (GCNH), a simple yet effective GNN architecture applicable to both heterophilous and homophilous scenarios.
arXiv Detail & Related papers (2023-04-21T11:26:24Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - Node Feature Extraction by Self-Supervised Multi-scale Neighborhood
Prediction [123.20238648121445]
We propose a new self-supervised learning framework, Graph Information Aided Node feature exTraction (GIANT)
GIANT makes use of the eXtreme Multi-label Classification (XMC) formalism, which is crucial for fine-tuning the language model based on graph information.
We demonstrate the superior performance of GIANT over the standard GNN pipeline on Open Graph Benchmark datasets.
arXiv Detail & Related papers (2021-10-29T19:55:12Z) - New Benchmarks for Learning on Non-Homophilous Graphs [20.082182515715182]
We present a series of improved graph datasets with node label relationships that do not satisfy the homophily principle.
We also introduce a new measure of the presence or absence of homophily that is better suited than existing measures in different regimes.
arXiv Detail & Related papers (2021-04-03T13:45:06Z) - Scalable Graph Neural Networks for Heterogeneous Graphs [12.44278942365518]
Graph neural networks (GNNs) are a popular class of parametric model for learning over graph-structured data.
Recent work has argued that GNNs primarily use the graph for feature smoothing, and have shown competitive results on benchmark tasks.
In this work, we ask whether these results can be extended to heterogeneous graphs, which encode multiple types of relationship between different entities.
arXiv Detail & Related papers (2020-11-19T06:03:35Z) - Combining Label Propagation and Simple Models Out-performs Graph Neural
Networks [52.121819834353865]
We show that for many standard transductive node classification benchmarks, we can exceed or match the performance of state-of-the-art GNNs.
We call this overall procedure Correct and Smooth (C&S)
Our approach exceeds or nearly matches the performance of state-of-the-art GNNs on a wide variety of benchmarks.
arXiv Detail & Related papers (2020-10-27T02:10:52Z) - Graph Contrastive Learning with Augmentations [109.23158429991298]
We propose a graph contrastive learning (GraphCL) framework for learning unsupervised representations of graph data.
We show that our framework can produce graph representations of similar or better generalizability, transferrability, and robustness compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-10-22T20:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.