GraphViz2Vec: A Structure-aware Feature Generation Model to Improve
Classification in GNNs
- URL: http://arxiv.org/abs/2401.17178v1
- Date: Tue, 30 Jan 2024 17:11:04 GMT
- Title: GraphViz2Vec: A Structure-aware Feature Generation Model to Improve
Classification in GNNs
- Authors: Shraban Kumar Chatterjee, Suman Kundu
- Abstract summary: GNNs are widely used to solve various tasks including node classification and link prediction.
In this paper, we presented a novel feature extraction methodology GraphViz2Vec that can capture structural information of a node's local neighbourhood.
These initial embeddings helps existing models achieve state-of-the-art results in various classification tasks.
- Score: 2.0823678201707034
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: GNNs are widely used to solve various tasks including node classification and
link prediction. Most of the GNN architectures assume the initial embedding to
be random or generated from popular distributions. These initial embeddings
require multiple layers of transformation to converge into a meaningful latent
representation. While number of layers allow accumulation of larger
neighbourhood of a node it also introduce the problem of over-smoothing. In
addition, GNNs are inept at representing structural information. For example,
the output embedding of a node does not capture its triangles participation. In
this paper, we presented a novel feature extraction methodology GraphViz2Vec
that can capture the structural information of a node's local neighbourhood to
create meaningful initial embeddings for a GNN model. These initial embeddings
helps existing models achieve state-of-the-art results in various
classification tasks. Further, these initial embeddings help the model to
produce desired results with only two layers which in turn reduce the problem
of over-smoothing. The initial encoding of a node is obtained from an image
classification model trained on multiple energy diagrams of its local
neighbourhood. These energy diagrams are generated with the induced sub-graph
of the nodes traversed by multiple random walks. The generated encodings
increase the performance of existing models on classification tasks (with a
mean increase of $4.65\%$ and $2.58\%$ for the node and link classification
tasks, respectively), with some models achieving state-of-the-art results.
Related papers
- Tackling Oversmoothing in GNN via Graph Sparsification: A Truss-based Approach [1.4854797901022863]
We propose a novel and flexible truss-based graph sparsification model that prunes edges from dense regions of the graph.
We then utilize our sparsification model in the state-of-the-art baseline GNNs and pooling models, such as GIN, SAGPool, GMT, DiffPool, MinCutPool, HGP-SL, DMonPool, and AdamGNN.
arXiv Detail & Related papers (2024-07-16T17:21:36Z) - Degree-based stratification of nodes in Graph Neural Networks [66.17149106033126]
We modify the Graph Neural Network (GNN) architecture so that the weight matrices are learned, separately, for the nodes in each group.
This simple-to-implement modification seems to improve performance across datasets and GNN methods.
arXiv Detail & Related papers (2023-12-16T14:09:23Z) - Content Augmented Graph Neural Networks [0.824969449883056]
We propose augmenting nodes' embeddings by embeddings generated from their content, at higher GNN layers.
We suggest methods such as using an auto-encoder or building a content graph, to generate content embeddings.
arXiv Detail & Related papers (2023-11-21T17:30:57Z) - NDGGNET-A Node Independent Gate based Graph Neural Networks [6.155450481110693]
For nodes with sparse connectivity, it is difficult to obtain enough information through a single GNN layer.
In this thesis, we define a novel framework that allows the normal GNN model to accommodate more layers.
Experimental results show that our proposed model can effectively increase the model depth and perform well on several datasets.
arXiv Detail & Related papers (2022-05-11T08:51:04Z) - A Variational Edge Partition Model for Supervised Graph Representation
Learning [51.30365677476971]
This paper introduces a graph generative process to model how the observed edges are generated by aggregating the node interactions over a set of overlapping node communities.
We partition each edge into the summation of multiple community-specific weighted edges and use them to define community-specific GNNs.
A variational inference framework is proposed to jointly learn a GNN based inference network that partitions the edges into different communities, these community-specific GNNs, and a GNN based predictor that combines community-specific GNNs for the end classification task.
arXiv Detail & Related papers (2022-02-07T14:37:50Z) - Feature Correlation Aggregation: on the Path to Better Graph Neural
Networks [37.79964911718766]
Prior to the introduction of Graph Neural Networks (GNNs), modeling and analyzing irregular data, particularly graphs, was thought to be the Achilles' heel of deep learning.
This paper introduces a central node permutation variant function through a frustratingly simple and innocent-looking modification to the core operation of a GNN.
A tangible boost in performance of the model is observed where the model surpasses previous state-of-the-art results by a significant margin while employing fewer parameters.
arXiv Detail & Related papers (2021-09-20T05:04:26Z) - Position-based Hash Embeddings For Scaling Graph Neural Networks [8.87527266373087]
Graph Neural Networks (GNNs) compute node representations by taking into account the topology of the node's ego-network and the features of the ego-network's nodes.
When the nodes do not have high-quality features, GNNs learn an embedding layer to compute node embeddings and use them as input features.
To reduce the memory associated with this embedding layer, hashing-based approaches, commonly used in applications like NLP and recommender systems, can potentially be used.
We present approaches that take advantage of the nodes' position in the graph to dramatically reduce the memory required.
arXiv Detail & Related papers (2021-08-31T22:42:25Z) - Explicit Pairwise Factorized Graph Neural Network for Semi-Supervised
Node Classification [59.06717774425588]
We propose the Explicit Pairwise Factorized Graph Neural Network (EPFGNN), which models the whole graph as a partially observed Markov Random Field.
It contains explicit pairwise factors to model output-output relations and uses a GNN backbone to model input-output relations.
We conduct experiments on various datasets, which shows that our model can effectively improve the performance for semi-supervised node classification on graphs.
arXiv Detail & Related papers (2021-07-27T19:47:53Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Distance Encoding: Design Provably More Powerful Neural Networks for
Graph Representation Learning [63.97983530843762]
Graph Neural Networks (GNNs) have achieved great success in graph representation learning.
GNNs generate identical representations for graph substructures that may in fact be very different.
More powerful GNNs, proposed recently by mimicking higher-order tests, are inefficient as they cannot sparsity of underlying graph structure.
We propose Distance Depiction (DE) as a new class of graph representation learning.
arXiv Detail & Related papers (2020-08-31T23:15:40Z) - Towards Deeper Graph Neural Networks with Differentiable Group
Normalization [61.20639338417576]
Graph neural networks (GNNs) learn the representation of a node by aggregating its neighbors.
Over-smoothing is one of the key issues which limit the performance of GNNs as the number of layers increases.
We introduce two over-smoothing metrics and a novel technique, i.e., differentiable group normalization (DGN)
arXiv Detail & Related papers (2020-06-12T07:18:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.