Network Representation Learning: From Preprocessing, Feature Extraction
to Node Embedding
- URL: http://arxiv.org/abs/2110.07582v1
- Date: Thu, 14 Oct 2021 17:46:37 GMT
- Title: Network Representation Learning: From Preprocessing, Feature Extraction
to Node Embedding
- Authors: Jingya Zhou, Ling Liu, Wenqi Wei, Jianxi Fan
- Abstract summary: Network representation learning (NRL) advances the conventional graph mining of social networks, knowledge graphs, and complex biomedical and physics information networks.
This survey paper reviews the design principles and the different node embedding techniques for network representation learning over homogeneous networks.
- Score: 9.844802841686105
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Network representation learning (NRL) advances the conventional graph mining
of social networks, knowledge graphs, and complex biomedical and physics
information networks. Over dozens of network representation learning algorithms
have been reported in the literature. Most of them focus on learning node
embeddings for homogeneous networks, but they differ in the specific encoding
schemes and specific types of node semantics captured and used for learning
node embedding. This survey paper reviews the design principles and the
different node embedding techniques for network representation learning over
homogeneous networks. To facilitate the comparison of different node embedding
algorithms, we introduce a unified reference framework to divide and generalize
the node embedding learning process on a given network into preprocessing
steps, node feature extraction steps and node embedding model training for a
NRL task such as link prediction and node clustering. With this unifying
reference framework, we highlight the representative methods, models, and
techniques used at different stages of the node embedding model learning
process. This survey not only helps researchers and practitioners to gain an
in-depth understanding of different network representation learning techniques
but also provides practical guidelines for designing and developing the next
generation of network representation learning algorithms and systems.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - GNN-LoFI: a Novel Graph Neural Network through Localized Feature-based
Histogram Intersection [51.608147732998994]
Graph neural networks are increasingly becoming the framework of choice for graph-based machine learning.
We propose a new graph neural network architecture that substitutes classical message passing with an analysis of the local distribution of node features.
arXiv Detail & Related papers (2024-01-17T13:04:23Z) - Associative Learning for Network Embedding [20.873120242498292]
We introduce a network embedding method from a new perspective.
Our network learns associations between the content of each node and that node's neighbors.
Our proposed method is evaluated on different downstream tasks such as node classification and linkage prediction.
arXiv Detail & Related papers (2022-08-30T16:35:45Z) - Topic-aware latent models for representation learning on networks [5.304857921982132]
We introduce TNE, a generic framework to enhance the embeddings of nodes acquired by means of random walk-based approaches with topic-based information.
We evaluate our methodology in two downstream tasks: node classification and link prediction.
arXiv Detail & Related papers (2021-11-10T08:52:52Z) - Temporal Graph Network Embedding with Causal Anonymous Walks
Representations [54.05212871508062]
We propose a novel approach for dynamic network representation learning based on Temporal Graph Network.
For evaluation, we provide a benchmark pipeline for the evaluation of temporal network embeddings.
We show the applicability and superior performance of our model in the real-world downstream graph machine learning task provided by one of the top European banks.
arXiv Detail & Related papers (2021-08-19T15:39:52Z) - Node2Seq: Towards Trainable Convolutions in Graph Neural Networks [59.378148590027735]
We propose a graph network layer, known as Node2Seq, to learn node embeddings with explicitly trainable weights for different neighboring nodes.
For a target node, our method sorts its neighboring nodes via attention mechanism and then employs 1D convolutional neural networks (CNNs) to enable explicit weights for information aggregation.
In addition, we propose to incorporate non-local information for feature learning in an adaptive manner based on the attention scores.
arXiv Detail & Related papers (2021-01-06T03:05:37Z) - DINE: A Framework for Deep Incomplete Network Embedding [33.97952453310253]
We propose a Deep Incomplete Network Embedding method, namely DINE.
We first complete the missing part including both nodes and edges in a partially observable network by using the expectation-maximization framework.
We evaluate DINE over three networks on multi-label classification and link prediction tasks.
arXiv Detail & Related papers (2020-08-09T04:59:35Z) - Modeling Dynamic Heterogeneous Network for Link Prediction using
Hierarchical Attention with Temporal RNN [16.362525151483084]
We propose a novel dynamic heterogeneous network embedding method, termed as DyHATR.
It uses hierarchical attention to learn heterogeneous information and incorporates recurrent neural networks with temporal attention to capture evolutionary patterns.
We benchmark our method on four real-world datasets for the task of link prediction.
arXiv Detail & Related papers (2020-04-01T17:16:47Z) - Temporal Network Representation Learning via Historical Neighborhoods
Aggregation [28.397309507168128]
We propose the Embedding via Historical Neighborhoods Aggregation (EHNA) algorithm.
We first propose a temporal random walk that can identify relevant nodes in historical neighborhoods.
Then we apply a deep learning model which uses a custom attention mechanism to induce node embeddings.
arXiv Detail & Related papers (2020-03-30T04:18:48Z) - Progressive Graph Convolutional Networks for Semi-Supervised Node
Classification [97.14064057840089]
Graph convolutional networks have been successful in addressing graph-based tasks such as semi-supervised node classification.
We propose a method to automatically build compact and task-specific graph convolutional networks.
arXiv Detail & Related papers (2020-03-27T08:32:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.