Local Graph Embeddings Based on Neighbors Degree Frequency of Nodes
- URL: http://arxiv.org/abs/2208.00152v1
- Date: Sat, 30 Jul 2022 07:07:30 GMT
- Title: Local Graph Embeddings Based on Neighbors Degree Frequency of Nodes
- Authors: Vahid Shirbisheh
- Abstract summary: We propose a strategy for graph machine learning and network analysis by defining certain local features and vector representations of nodes.
By extending the notion of the degree of a node via Breath-First Search, a general family of bf centrality functions is defined.
We show that centrality and closeness can be learned by applying deep learning to these local features.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a local-to-global strategy for graph machine learning and network
analysis by defining certain local features and vector representations of nodes
and then using them to learn globally defined metrics and properties of the
nodes by means of deep neural networks. By extending the notion of the degree
of a node via Breath-First Search, a general family of {\bf parametric
centrality functions} is defined which are able to reveal the importance of
nodes. We introduce the {\bf neighbors degree frequency (NDF)}, as a locally
defined embedding of nodes of undirected graphs into euclidean spaces. This
gives rise to a vectorized labeling of nodes which encodes the structure of
local neighborhoods of nodes and can be used for graph isomorphism testing. We
add flexibility to our construction so that it can handle dynamic graphs as
well. Afterwards, the Breadth-First Search is used to extend NDF vector
representations into two different matrix representations of nodes which
contain higher order information about the neighborhoods of nodes. Our matrix
representations of nodes provide us with a new way of visualizing the shape of
the neighborhood of a node. Furthermore, we use these matrix representations to
obtain feature vectors, which are suitable for typical deep learning
algorithms. To demonstrate these node embeddings actually contain some
information about the nodes, in a series of examples, we show that PageRank and
closeness centrality can be learned by applying deep learning to these local
features. Our constructions are flexible enough to handle evolving graphs.
Finally, we explain how to adapt our constructions for directed graphs.
Related papers
- Saliency-Aware Regularized Graph Neural Network [39.82009838086267]
We propose the Saliency-Aware Regularized Graph Neural Network (SAR-GNN) for graph classification.
We first estimate the global node saliency by measuring the semantic similarity between the compact graph representation and node features.
Then the learned saliency distribution is leveraged to regularize the neighborhood aggregation of the backbone.
arXiv Detail & Related papers (2024-01-01T13:44:16Z) - GraphRARE: Reinforcement Learning Enhanced Graph Neural Network with Relative Entropy [21.553180564868306]
GraphRARE is a framework built upon node relative entropy and deep reinforcement learning.
An innovative node relative entropy is used to measure mutual information between node pairs.
A deep reinforcement learning-based algorithm is developed to optimize the graph topology.
arXiv Detail & Related papers (2023-12-15T11:30:18Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Graph Neural Networks with Feature and Structure Aware Random Walk [7.143879014059894]
We show that in typical heterphilous graphs, the edges may be directed, and whether to treat the edges as is or simply make them undirected greatly affects the performance of the GNN models.
We develop a model that adaptively learns the directionality of the graph, and exploits the underlying long-distance correlations between nodes.
arXiv Detail & Related papers (2021-11-19T08:54:21Z) - Graph Neural Networks with Learnable Structural and Positional
Representations [83.24058411666483]
A major issue with arbitrary graphs is the absence of canonical positional information of nodes.
We introduce Positional nodes (PE) of nodes, and inject it into the input layer, like in Transformers.
We observe a performance increase for molecular datasets, from 2.87% up to 64.14% when considering learnable PE for both GNN classes.
arXiv Detail & Related papers (2021-10-15T05:59:15Z) - COLOGNE: Coordinated Local Graph Neighborhood Sampling [1.6498361958317633]
replacing discrete unordered objects such as graph nodes by real-valued vectors is at the heart of many approaches to learning from graph data.
We address the problem of learning discrete node embeddings such that the coordinates of the node vector representations are graph nodes.
This opens the door to designing interpretable machine learning algorithms for graphs as all attributes originally present in the nodes are preserved.
arXiv Detail & Related papers (2021-02-09T11:39:06Z) - Node2Seq: Towards Trainable Convolutions in Graph Neural Networks [59.378148590027735]
We propose a graph network layer, known as Node2Seq, to learn node embeddings with explicitly trainable weights for different neighboring nodes.
For a target node, our method sorts its neighboring nodes via attention mechanism and then employs 1D convolutional neural networks (CNNs) to enable explicit weights for information aggregation.
In addition, we propose to incorporate non-local information for feature learning in an adaptive manner based on the attention scores.
arXiv Detail & Related papers (2021-01-06T03:05:37Z) - node2coords: Graph Representation Learning with Wasserstein Barycenters [59.07120857271367]
We introduce node2coords, a representation learning algorithm for graphs.
It learns simultaneously a low-dimensional space and coordinates for the nodes in that space.
Experimental results demonstrate that the representations learned with node2coords are interpretable.
arXiv Detail & Related papers (2020-07-31T13:14:25Z) - Graph Neural Networks with Composite Kernels [60.81504431653264]
We re-interpret node aggregation from the perspective of kernel weighting.
We present a framework to consider feature similarity in an aggregation scheme.
We propose feature aggregation as the composition of the original neighbor-based kernel and a learnable kernel to encode feature similarities in a feature space.
arXiv Detail & Related papers (2020-05-16T04:44:29Z) - Graph Inference Learning for Semi-supervised Classification [50.55765399527556]
We propose a Graph Inference Learning framework to boost the performance of semi-supervised node classification.
For learning the inference process, we introduce meta-optimization on structure relations from training nodes to validation nodes.
Comprehensive evaluations on four benchmark datasets demonstrate the superiority of our proposed GIL when compared against state-of-the-art methods.
arXiv Detail & Related papers (2020-01-17T02:52:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.