Orthogonal Graph Neural Networks
- URL: http://arxiv.org/abs/2109.11338v1
- Date: Thu, 23 Sep 2021 12:39:01 GMT
- Title: Orthogonal Graph Neural Networks
- Authors: Kai Guo, Kaixiong Zhou, Xia Hu, Yu Li, Yi Chang, Xin Wang
- Abstract summary: Graph neural networks (GNNs) have received tremendous attention due to their superiority in learning node representations.
stacking more convolutional layers significantly decreases the performance of GNNs.
We propose a novel Ortho-GConv, which could generally augment the existing GNN backbones to stabilize the model training and improve the model's generalization performance.
- Score: 53.466187667936026
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph neural networks (GNNs) have received tremendous attention due to their
superiority in learning node representations. These models rely on message
passing and feature transformation functions to encode the structural and
feature information from neighbors. However, stacking more convolutional layers
significantly decreases the performance of GNNs. Most recent studies attribute
this limitation to the over-smoothing issue, where node embeddings converge to
indistinguishable vectors. Through a number of experimental observations, we
argue that the main factor degrading the performance is the unstable forward
normalization and backward gradient resulted from the improper design of the
feature transformation, especially for shallow GNNs where the over-smoothing
has not happened. Therefore, we propose a novel orthogonal feature
transformation, named Ortho-GConv, which could generally augment the existing
GNN backbones to stabilize the model training and improve the model's
generalization performance. Specifically, we maintain the orthogonality of the
feature transformation comprehensively from three perspectives, namely hybrid
weight initialization, orthogonal transformation, and orthogonal
regularization. By equipping the existing GNNs (e.g. GCN, JKNet, GCNII) with
Ortho-GConv, we demonstrate the generality of the orthogonal feature
transformation to enable stable training, and show its effectiveness for node
and graph classification tasks.
Related papers
- Learning to Control the Smoothness of Graph Convolutional Network Features [9.949988676706418]
We propose a new strategy to let graph convolutional network (GCN) learn node features with a desired smoothness.
Our approach has three key steps: We establish a geometric relationship between the input and output of ReLU or leaky ReLU.
Building on our geometric insights, we augment the message-passing process of graph convolutional layers with a learnable term to modulate the smoothness of node features.
arXiv Detail & Related papers (2024-10-18T16:57:27Z) - Degree-based stratification of nodes in Graph Neural Networks [66.17149106033126]
We modify the Graph Neural Network (GNN) architecture so that the weight matrices are learned, separately, for the nodes in each group.
This simple-to-implement modification seems to improve performance across datasets and GNN methods.
arXiv Detail & Related papers (2023-12-16T14:09:23Z) - On the Initialization of Graph Neural Networks [10.153841274798829]
We analyze the variance of forward and backward propagation across Graph Neural Networks layers.
We propose a new method for Variance Instability Reduction within GNN Optimization (Virgo)
We conduct comprehensive experiments on 15 datasets to show that Virgo can lead to superior model performance.
arXiv Detail & Related papers (2023-12-05T09:55:49Z) - OrthoReg: Improving Graph-regularized MLPs via Orthogonality
Regularization [66.30021126251725]
Graph Neural Networks (GNNs) are currently dominating in modeling graphstructure data.
Graph-regularized networks (GR-MLPs) implicitly inject the graph structure information into model weights, while their performance can hardly match that of GNNs in most tasks.
We show that GR-MLPs suffer from dimensional collapse, a phenomenon in which the largest a few eigenvalues dominate the embedding space.
We propose OrthoReg, a novel GR-MLP model to mitigate the dimensional collapse issue.
arXiv Detail & Related papers (2023-01-31T21:20:48Z) - ResNorm: Tackling Long-tailed Degree Distribution Issue in Graph Neural
Networks via Normalization [80.90206641975375]
This paper focuses on improving the performance of GNNs via normalization.
By studying the long-tailed distribution of node degrees in the graph, we propose a novel normalization method for GNNs.
The $scale$ operation of ResNorm reshapes the node-wise standard deviation (NStd) distribution so as to improve the accuracy of tail nodes.
arXiv Detail & Related papers (2022-06-16T13:49:09Z) - Revisiting Transformation Invariant Geometric Deep Learning: Are Initial
Representations All You Need? [80.86819657126041]
We show that transformation-invariant and distance-preserving initial representations are sufficient to achieve transformation invariance.
Specifically, we realize transformation-invariant and distance-preserving initial point representations by modifying multi-dimensional scaling.
We prove that TinvNN can strictly guarantee transformation invariance, being general and flexible enough to be combined with the existing neural networks.
arXiv Detail & Related papers (2021-12-23T03:52:33Z) - Feature Correlation Aggregation: on the Path to Better Graph Neural
Networks [37.79964911718766]
Prior to the introduction of Graph Neural Networks (GNNs), modeling and analyzing irregular data, particularly graphs, was thought to be the Achilles' heel of deep learning.
This paper introduces a central node permutation variant function through a frustratingly simple and innocent-looking modification to the core operation of a GNN.
A tangible boost in performance of the model is observed where the model surpasses previous state-of-the-art results by a significant margin while employing fewer parameters.
arXiv Detail & Related papers (2021-09-20T05:04:26Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.