DRGCN: Dynamic Evolving Initial Residual for Deep Graph Convolutional
Networks
- URL: http://arxiv.org/abs/2302.05083v1
- Date: Fri, 10 Feb 2023 06:57:12 GMT
- Title: DRGCN: Dynamic Evolving Initial Residual for Deep Graph Convolutional
Networks
- Authors: Lei Zhang, Xiaodong Yan, Jianshan He, Ruopeng Li, Wei Chu
- Abstract summary: We propose a novel model called Dynamic evolving initial Residual Graph Convolutional Network (DRGCN)
Our experimental results show that our model effectively relieves the problem of over-smoothing in deep GCNs.
Our model reaches new SOTA results on the large-scale ogbn-arxiv dataset of Open Graph Benchmark (OGB)
- Score: 19.483662490506646
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph convolutional networks (GCNs) have been proved to be very practical to
handle various graph-related tasks. It has attracted considerable research
interest to study deep GCNs, due to their potential superior performance
compared with shallow ones. However, simply increasing network depth will, on
the contrary, hurt the performance due to the over-smoothing problem. Adding
residual connection is proved to be effective for learning deep convolutional
neural networks (deep CNNs), it is not trivial when applied to deep GCNs.
Recent works proposed an initial residual mechanism that did alleviate the
over-smoothing problem in deep GCNs. However, according to our study, their
algorithms are quite sensitive to different datasets. In their setting, the
personalization (dynamic) and correlation (evolving) of how residual applies
are ignored. To this end, we propose a novel model called Dynamic evolving
initial Residual Graph Convolutional Network (DRGCN). Firstly, we use a dynamic
block for each node to adaptively fetch information from the initial
representation. Secondly, we use an evolving block to model the residual
evolving pattern between layers. Our experimental results show that our model
effectively relieves the problem of over-smoothing in deep GCNs and outperforms
the state-of-the-art (SOTA) methods on various benchmark datasets. Moreover, we
develop a mini-batch version of DRGCN which can be applied to large-scale data.
Coupling with several fair training techniques, our model reaches new SOTA
results on the large-scale ogbn-arxiv dataset of Open Graph Benchmark (OGB).
Our reproducible code is available on GitHub.
Related papers
- An Introduction to Robust Graph Convolutional Networks [71.68610791161355]
We propose a novel Robust Graph Convolutional Neural Networks for possible erroneous single-view or multi-view data.
By incorporating an extra layers via Autoencoders into traditional graph convolutional networks, we characterize and handle typical error models explicitly.
arXiv Detail & Related papers (2021-03-27T04:47:59Z) - Overcoming Catastrophic Forgetting in Graph Neural Networks [50.900153089330175]
Catastrophic forgetting refers to the tendency that a neural network "forgets" the previous learned knowledge upon learning new tasks.
We propose a novel scheme dedicated to overcoming this problem and hence strengthen continual learning in graph neural networks (GNNs)
At the heart of our approach is a generic module, termed as topology-aware weight preserving(TWP)
arXiv Detail & Related papers (2020-12-10T22:30:25Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - GraphMDN: Leveraging graph structure and deep learning to solve inverse
problems [0.0]
We develop a Graph Mixture Density Network (GraphMDN), which combines graph neural networks with mixture density network (MDN) outputs.
GraphMDNs excel on regression tasks wherein the data are graph structured, and target statistics are better represented by mixtures of densities.
arXiv Detail & Related papers (2020-10-26T15:44:22Z) - Graph Convolutional Networks for Graphs Containing Missing Features [5.426650977249329]
We propose an approach that adapts Graph Convolutional Network (GCN) to graphs containing missing features.
In contrast to traditional strategy, our approach integrates the processing of missing features and graph learning within the same neural network architecture.
We demonstrate through extensive experiments that our approach significantly outperforms the imputation-based methods in node classification and link prediction tasks.
arXiv Detail & Related papers (2020-07-09T06:47:21Z) - Simple and Deep Graph Convolutional Networks [63.76221532439285]
Graph convolutional networks (GCNs) are a powerful deep learning approach for graph-structured data.
Despite their success, most of the current GCN models are shallow, due to the em over-smoothing problem.
We propose the GCNII, an extension of the vanilla GCN model with two simple yet effective techniques.
arXiv Detail & Related papers (2020-07-04T16:18:06Z) - Understanding and Resolving Performance Degradation in Graph
Convolutional Networks [105.14867349802898]
Graph Convolutional Network (GCN) stacks several layers and in each layer performs a PROPagation operation (PROP) and a TRANsformation operation (TRAN) for learning node representations over graph-structured data.
GCNs tend to suffer performance drop when the model gets deep.
We study performance degradation of GCNs by experimentally examining how stacking only TRANs or PROPs works.
arXiv Detail & Related papers (2020-06-12T12:12:12Z) - Revisiting Graph based Collaborative Filtering: A Linear Residual Graph
Convolutional Network Approach [55.44107800525776]
Graph Convolutional Networks (GCNs) are state-of-the-art graph based representation learning models.
In this paper, we revisit GCN based Collaborative Filtering (CF) based Recommender Systems (RS)
We show that removing non-linearities would enhance recommendation performance, consistent with the theories in simple graph convolutional networks.
We propose a residual network structure that is specifically designed for CF with user-item interaction modeling.
arXiv Detail & Related papers (2020-01-28T04:41:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.