Wiki-CS: A Wikipedia-Based Benchmark for Graph Neural Networks
- URL: http://arxiv.org/abs/2007.02901v2
- Date: Sun, 9 Jan 2022 23:19:38 GMT
- Title: Wiki-CS: A Wikipedia-Based Benchmark for Graph Neural Networks
- Authors: P\'eter Mernyei, C\u{a}t\u{a}lina Cangea
- Abstract summary: We present Wiki-CS, a novel dataset derived from Wikipedia for benchmarking Graph Neural Networks.
The dataset consists of nodes corresponding to Computer Science articles, with edges based on hyperlinks and 10 classes representing different branches of the field.
We use the dataset to evaluate semi-supervised node classification and single-relation link prediction models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present Wiki-CS, a novel dataset derived from Wikipedia for benchmarking
Graph Neural Networks. The dataset consists of nodes corresponding to Computer
Science articles, with edges based on hyperlinks and 10 classes representing
different branches of the field. We use the dataset to evaluate semi-supervised
node classification and single-relation link prediction models. Our experiments
show that these methods perform well on a new domain, with structural
properties different from earlier benchmarks. The dataset is publicly
available, along with the implementation of the data pipeline and the benchmark
experiments, at https://github.com/pmernyei/wiki-cs-dataset .
Related papers
- Temporal Graph Benchmark for Machine Learning on Temporal Graphs [54.52243310226456]
Temporal Graph Benchmark (TGB) is a collection of challenging and diverse benchmark datasets.
We benchmark each dataset and find that the performance of common models can vary drastically across datasets.
TGB provides an automated machine learning pipeline for reproducible and accessible temporal graph research.
arXiv Detail & Related papers (2023-07-03T13:58:20Z) - Graphtester: Exploring Theoretical Boundaries of GNNs on Graph Datasets [10.590698823137755]
We provide a new tool called Graphtester for a comprehensive analysis of the theoretical capabilities of GNNs for various datasets, tasks, and scores.
We use Graphtester to analyze over 40 different graph datasets, determining upper bounds on the performance of various GNNs based on the number of layers.
We show that the tool can also be used for Graph Transformers using positional node encodings, thereby expanding its scope.
arXiv Detail & Related papers (2023-06-30T08:53:23Z) - A Robust Stacking Framework for Training Deep Graph Models with
Multifaceted Node Features [61.92791503017341]
Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data.
The best models for such data types in most standard supervised learning settings with IID (non-graph) data are not easily incorporated into a GNN.
Here we propose a robust stacking framework that fuses graph-aware propagation with arbitrary models intended for IID data.
arXiv Detail & Related papers (2022-06-16T22:46:33Z) - SHGNN: Structure-Aware Heterogeneous Graph Neural Network [77.78459918119536]
This paper proposes a novel Structure-Aware Heterogeneous Graph Neural Network (SHGNN) to address the above limitations.
We first utilize a feature propagation module to capture the local structure information of intermediate nodes in the meta-path.
Next, we use a tree-attention aggregator to incorporate the graph structure information into the aggregation module on the meta-path.
Finally, we leverage a meta-path aggregator to fuse the information aggregated from different meta-paths.
arXiv Detail & Related papers (2021-12-12T14:18:18Z) - Temporal Graph Network Embedding with Causal Anonymous Walks
Representations [54.05212871508062]
We propose a novel approach for dynamic network representation learning based on Temporal Graph Network.
For evaluation, we provide a benchmark pipeline for the evaluation of temporal network embeddings.
We show the applicability and superior performance of our model in the real-world downstream graph machine learning task provided by one of the top European banks.
arXiv Detail & Related papers (2021-08-19T15:39:52Z) - A pipeline for fair comparison of graph neural networks in node
classification tasks [4.418753792543564]
Graph neural networks (GNNs) have been investigated for potential applicability in multiple fields that employ graph data.
There are no standard training settings to ensure fair comparisons among new methods.
We introduce a standard, reproducible benchmark to which the same training settings can be applied for node classification.
arXiv Detail & Related papers (2020-12-19T07:43:05Z) - TUDataset: A collection of benchmark datasets for learning with graphs [21.16723995518478]
We introduce the TUDataset for graph classification and regression.
The collection consists of over 120 datasets of varying sizes from a wide range of applications.
All datasets are available at www.graphlearning.io.
arXiv Detail & Related papers (2020-07-16T21:46:33Z) - Sequential Graph Convolutional Network for Active Learning [53.99104862192055]
We propose a novel pool-based Active Learning framework constructed on a sequential Graph Convolution Network (GCN)
With a small number of randomly sampled images as seed labelled examples, we learn the parameters of the graph to distinguish labelled vs unlabelled nodes.
We exploit these characteristics of GCN to select the unlabelled examples which are sufficiently different from labelled ones.
arXiv Detail & Related papers (2020-06-18T00:55:10Z) - Convolutional Kernel Networks for Graph-Structured Data [37.13712126432493]
We introduce a family of multilayer graph kernels and establish new links between graph convolutional neural networks and kernel methods.
Our approach generalizes convolutional kernel networks to graph-structured data, by representing graphs as a sequence of kernel feature maps.
Our model can also be trained end-to-end on large-scale data, leading to new types of graph convolutional neural networks.
arXiv Detail & Related papers (2020-03-11T09:44:03Z) - Graph Inference Learning for Semi-supervised Classification [50.55765399527556]
We propose a Graph Inference Learning framework to boost the performance of semi-supervised node classification.
For learning the inference process, we introduce meta-optimization on structure relations from training nodes to validation nodes.
Comprehensive evaluations on four benchmark datasets demonstrate the superiority of our proposed GIL when compared against state-of-the-art methods.
arXiv Detail & Related papers (2020-01-17T02:52:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.