Quadratic GCN for Graph Classification
- URL: http://arxiv.org/abs/2104.06750v1
- Date: Wed, 14 Apr 2021 10:01:09 GMT
- Title: Quadratic GCN for Graph Classification
- Authors: Omer Nagar, Shoval Frydman, Ori Hochman and Yoram Louzoun
- Abstract summary: Graph Convolutional Networks (GCNs) have been extensively used to classify vertices in graphs.
GCNs have been extended to graph classification tasks (GCT)
We propose a novel solution combining GCN, methods from knowledge graphs, and a new self-regularized activation function.
- Score: 1.1470070927586016
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Graph Convolutional Networks (GCNs) have been extensively used to classify
vertices in graphs and have been shown to outperform other vertex
classification methods. GCNs have been extended to graph classification tasks
(GCT). In GCT, graphs with different numbers of edges and vertices belong to
different classes, and one attempts to predict the graph class. GCN based GCT
have mostly used pooling and attention-based models. The accuracy of existing
GCT methods is still limited. We here propose a novel solution combining GCN,
methods from knowledge graphs, and a new self-regularized activation function
to significantly improve the accuracy of the GCN based GCT. We present
quadratic GCN (QGCN) - A GCN formalism with a quadratic layer. Such a layer
produces an output with fixed dimensions, independent of the graph vertex
number. We applied this method to a wide range of graph classification
problems, and show that when using a self regularized activation function, QGCN
outperforms the state of the art methods for all graph classification tasks
tested with or without external input on each graph. The code for QGCN is
available at: https://github.com/Unknown-Data/QGCN .
Related papers
- Spectral Greedy Coresets for Graph Neural Networks [61.24300262316091]
The ubiquity of large-scale graphs in node-classification tasks hinders the real-world applications of Graph Neural Networks (GNNs)
This paper studies graph coresets for GNNs and avoids the interdependence issue by selecting ego-graphs based on their spectral embeddings.
Our spectral greedy graph coreset (SGGC) scales to graphs with millions of nodes, obviates the need for model pre-training, and applies to low-homophily graphs.
arXiv Detail & Related papers (2024-05-27T17:52:12Z) - Graph Convolutional Network For Semi-supervised Node Classification With Subgraph Sketching [0.27624021966289597]
We propose the Graph-Learning-Dual Graph Convolutional Neural Network called GLDGCN.
We apply GLDGCN to the semi-supervised node classification task.
Compared with the baseline methods, we achieve higher classification accuracy on three citation networks.
arXiv Detail & Related papers (2024-04-19T09:08:12Z) - AnchorGAE: General Data Clustering via $O(n)$ Bipartite Graph
Convolution [79.44066256794187]
We show how to convert a non-graph dataset into a graph by introducing the generative graph model, which is used to build graph convolution networks (GCNs)
A bipartite graph constructed by anchors is updated dynamically to exploit the high-level information behind data.
We theoretically prove that the simple update will lead to degeneration and a specific strategy is accordingly designed.
arXiv Detail & Related papers (2021-11-12T07:08:13Z) - Graph Convolutional Networks for Graphs Containing Missing Features [5.426650977249329]
We propose an approach that adapts Graph Convolutional Network (GCN) to graphs containing missing features.
In contrast to traditional strategy, our approach integrates the processing of missing features and graph learning within the same neural network architecture.
We demonstrate through extensive experiments that our approach significantly outperforms the imputation-based methods in node classification and link prediction tasks.
arXiv Detail & Related papers (2020-07-09T06:47:21Z) - Investigating and Mitigating Degree-Related Biases in Graph
Convolutional Networks [62.8504260693664]
Graph Convolutional Networks (GCNs) show promising results for semisupervised learning tasks on graphs.
In this paper, we analyze GCNs in regard to the node degree distribution.
We develop a novel Self-Supervised DegreeSpecific GCN (SL-DSGC) that mitigates the degree biases of GCNs.
arXiv Detail & Related papers (2020-06-28T16:26:47Z) - Connecting Graph Convolutional Networks and Graph-Regularized PCA [23.592657600394215]
Graph convolution operator of the GCN model is originally motivated from a localized first-order approximation of spectral graph convolutions.
We establish a textitmathematical connection between graph convolution and graph-regularized PCA (GPCA)
arXiv Detail & Related papers (2020-06-22T14:27:02Z) - DeeperGCN: All You Need to Train Deeper GCNs [66.64739331859226]
Graph Convolutional Networks (GCNs) have been drawing significant attention with the power of representation learning on graphs.
Unlike Convolutional Neural Networks (CNNs), which are able to take advantage of stacking very deep layers, GCNs suffer from vanishing gradient, over-smoothing and over-fitting issues when going deeper.
This paper proposes DeeperGCN that is capable of successfully and reliably training very deep GCNs.
arXiv Detail & Related papers (2020-06-13T23:00:22Z) - Directed Graph Convolutional Network [15.879411956536885]
We extend spectral-based graph convolution to directed graphs by using first- and second-order proximity.
A new GCN model, called DGCN, is then designed to learn representations on the directed graph.
arXiv Detail & Related papers (2020-04-29T06:19:10Z) - Cross-GCN: Enhancing Graph Convolutional Network with $k$-Order Feature
Interactions [153.6357310444093]
Graph Convolutional Network (GCN) is an emerging technique that performs learning and reasoning on graph data.
We argue that existing designs of GCN forgo modeling cross features, making GCN less effective for tasks or data where cross features are important.
We design a new operator named Cross-feature Graph Convolution, which explicitly models the arbitrary-order cross features with complexity linear to feature dimension and order size.
arXiv Detail & Related papers (2020-03-05T13:05:27Z) - Infinitely Wide Graph Convolutional Networks: Semi-supervised Learning
via Gaussian Processes [144.6048446370369]
Graph convolutional neural networks(GCNs) have recently demonstrated promising results on graph-based semi-supervised classification.
We propose a GP regression model via GCNs(GPGC) for graph-based semi-supervised learning.
We conduct extensive experiments to evaluate GPGC and demonstrate that it outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2020-02-26T10:02:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.