Designing the Topology of Graph Neural Networks: A Novel Feature Fusion
Perspective
- URL: http://arxiv.org/abs/2112.14531v1
- Date: Wed, 29 Dec 2021 13:06:12 GMT
- Title: Designing the Topology of Graph Neural Networks: A Novel Feature Fusion
Perspective
- Authors: Lanning Wei, Huan Zhao, Zhiqiang He
- Abstract summary: We learn to design the topology of GNNs in a novel feature fusion perspective which is dubbed F$2$GNN.
We develop a neural architecture search method on top of the unified framework which contains a set of selection and fusion operations.
The performance gains on eight real-world datasets demonstrate the effectiveness of F$2$GNN.
- Score: 12.363386808994079
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, Graph Neural Networks (GNNs) have shown superior performance
on diverse real-world applications. To improve the model capacity, besides
designing aggregation operations, GNN topology design is also very important.
In general, there are two mainstream GNN topology design manners. The first one
is to stack aggregation operations to obtain the higher-level features but
easily got performance drop as the network goes deeper. Secondly, the multiple
aggregation operations are utilized in each layer which provides adequate and
independent feature extraction stage on local neighbors while are costly to
obtain the higher-level information. To enjoy the benefits while alleviating
the corresponding deficiencies of these two manners, we learn to design the
topology of GNNs in a novel feature fusion perspective which is dubbed
F$^2$GNN. To be specific, we provide a feature fusion perspective in designing
GNN topology and propose a novel framework to unify the existing topology
designs with feature selection and fusion strategies. Then we develop a neural
architecture search method on top of the unified framework which contains a set
of selection and fusion operations in the search space and an improved
differentiable search algorithm. The performance gains on eight real-world
datasets demonstrate the effectiveness of F$^2$GNN. We further conduct
experiments to show that F$^2$GNN can improve the model capacity while
alleviating the deficiencies of existing GNN topology design manners,
especially alleviating the over-smoothing problem, by utilizing different
levels of features adaptively.
Related papers
- AGNN: Alternating Graph-Regularized Neural Networks to Alleviate
Over-Smoothing [29.618952407794776]
We propose an Alternating Graph-regularized Neural Network (AGNN) composed of Graph Convolutional Layer (GCL) and Graph Embedding Layer (GEL)
GEL is derived from the graph-regularized optimization containing Laplacian embedding term, which can alleviate the over-smoothing problem.
AGNN is evaluated via a large number of experiments including performance comparison with some multi-layer or multi-order graph neural networks.
arXiv Detail & Related papers (2023-04-14T09:20:03Z) - Automatic Relation-aware Graph Network Proliferation [182.30735195376792]
We propose Automatic Relation-aware Graph Network Proliferation (ARGNP) for efficiently searching GNNs.
These operations can extract hierarchical node/relational information and provide anisotropic guidance for message passing on a graph.
Experiments on six datasets for four graph learning tasks demonstrate that GNNs produced by our method are superior to the current state-of-the-art hand-crafted and search-based GNNs.
arXiv Detail & Related papers (2022-05-31T10:38:04Z) - Space4HGNN: A Novel, Modularized and Reproducible Platform to Evaluate
Heterogeneous Graph Neural Network [51.07168862821267]
We propose a unified framework covering most HGNNs, consisting of three components: heterogeneous linear transformation, heterogeneous graph transformation, and heterogeneous message passing layer.
We then build a platform Space4HGNN by defining a design space for HGNNs based on the unified framework, which offers modularized components, reproducible implementations, and standardized evaluation for HGNNs.
arXiv Detail & Related papers (2022-02-18T13:11:35Z) - Simplifying approach to Node Classification in Graph Neural Networks [7.057970273958933]
We decouple the node feature aggregation step and depth of graph neural network, and empirically analyze how different aggregated features play a role in prediction performance.
We show that not all features generated via aggregation steps are useful, and often using these less informative features can be detrimental to the performance of the GNN model.
We present a simple and shallow model, Feature Selection Graph Neural Network (FSGNN), and show empirically that the proposed model achieves comparable or even higher accuracy than state-of-the-art GNN models.
arXiv Detail & Related papers (2021-11-12T14:53:22Z) - ACE-HGNN: Adaptive Curvature Exploration Hyperbolic Graph Neural Network [72.16255675586089]
We propose an Adaptive Curvature Exploration Hyperbolic Graph NeuralNetwork named ACE-HGNN to adaptively learn the optimal curvature according to the input graph and downstream tasks.
Experiments on multiple real-world graph datasets demonstrate a significant and consistent performance improvement in model quality with competitive performance and good generalization ability.
arXiv Detail & Related papers (2021-10-15T07:18:57Z) - Search For Deep Graph Neural Networks [4.3002928862077825]
Current GNN-oriented NAS methods focus on the search for different layer aggregate components with shallow and simple architectures.
We propose a GNN generation pipeline with a novel two-stage search space, which aims at automatically generating high-performance.
Experiments on real-world datasets show that our generated GNN models outperforms existing manually designed and NAS-based ones.
arXiv Detail & Related papers (2021-09-21T09:24:59Z) - Edge-featured Graph Neural Architecture Search [131.4361207769865]
We propose Edge-featured Graph Neural Architecture Search to find the optimal GNN architecture.
Specifically, we design rich entity and edge updating operations to learn high-order representations.
We show EGNAS can search better GNNs with higher performance than current state-of-the-art human-designed and searched-based GNNs.
arXiv Detail & Related papers (2021-09-03T07:53:18Z) - Enhance Information Propagation for Graph Neural Network by
Heterogeneous Aggregations [7.3136594018091134]
Graph neural networks are emerging as continuation of deep learning success w.r.t. graph data.
We propose to enhance information propagation among GNN layers by combining heterogeneous aggregations.
We empirically validate the effectiveness of HAG-Net on a number of graph classification benchmarks.
arXiv Detail & Related papers (2021-02-08T08:57:56Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Policy-GNN: Aggregation Optimization for Graph Neural Networks [60.50932472042379]
Graph neural networks (GNNs) aim to model the local graph structures and capture the hierarchical patterns by aggregating the information from neighbors.
It is a challenging task to develop an effective aggregation strategy for each node, given complex graphs and sparse features.
We propose Policy-GNN, a meta-policy framework that models the sampling procedure and message passing of GNNs into a combined learning process.
arXiv Detail & Related papers (2020-06-26T17:03:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.