Graph-Level Embedding for Time-Evolving Graphs
- URL: http://arxiv.org/abs/2306.01012v1
- Date: Thu, 1 Jun 2023 01:50:37 GMT
- Title: Graph-Level Embedding for Time-Evolving Graphs
- Authors: Lili Wang, Chenghan Huang, Weicheng Ma, Xinyuan Cao, and Soroush
Vosoughi
- Abstract summary: Graph representation learning (also known as network embedding) has been extensively researched with varying levels of granularity.
We present a novel method for temporal graph-level embedding that addresses this gap.
- Score: 24.194795771873046
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Graph representation learning (also known as network embedding) has been
extensively researched with varying levels of granularity, ranging from nodes
to graphs. While most prior work in this area focuses on node-level
representation, limited research has been conducted on graph-level embedding,
particularly for dynamic or temporal networks. However, learning
low-dimensional graph-level representations for dynamic networks is critical
for various downstream graph retrieval tasks such as temporal graph similarity
ranking, temporal graph isomorphism, and anomaly detection. In this paper, we
present a novel method for temporal graph-level embedding that addresses this
gap. Our approach involves constructing a multilayer graph and using a modified
random walk with temporal backtracking to generate temporal contexts for the
graph's nodes. We then train a "document-level" language model on these
contexts to generate graph-level embeddings. We evaluate our proposed model on
five publicly available datasets for the task of temporal graph similarity
ranking, and our model outperforms baseline methods. Our experimental results
demonstrate the effectiveness of our method in generating graph-level
embeddings for dynamic networks.
Related papers
- Knowledge Probing for Graph Representation Learning [12.960185655357495]
We propose a novel graph probing framework (GraphProbe) to investigate and interpret whether the family of graph learning methods has encoded different levels of knowledge in graph representation learning.
Based on the intrinsic properties of graphs, we design three probes to systematically investigate the graph representation learning process from different perspectives.
We construct a thorough evaluation benchmark with nine representative graph learning methods from random walk based approaches, basic graph neural networks and self-supervised graph methods, and probe them on six benchmark datasets for node classification, link prediction and graph classification.
arXiv Detail & Related papers (2024-08-07T16:27:45Z) - Deep Temporal Graph Clustering [77.02070768950145]
We propose a general framework for deep Temporal Graph Clustering (GC)
GC introduces deep clustering techniques to suit the interaction sequence-based batch-processing pattern of temporal graphs.
Our framework can effectively improve the performance of existing temporal graph learning methods.
arXiv Detail & Related papers (2023-05-18T06:17:50Z) - State of the Art and Potentialities of Graph-level Learning [54.68482109186052]
Graph-level learning has been applied to many tasks including comparison, regression, classification, and more.
Traditional approaches to learning a set of graphs rely on hand-crafted features, such as substructures.
Deep learning has helped graph-level learning adapt to the growing scale of graphs by extracting features automatically and encoding graphs into low-dimensional representations.
arXiv Detail & Related papers (2023-01-14T09:15:49Z) - Towards Real-Time Temporal Graph Learning [10.647431919265346]
We propose an end-to-end graph learning pipeline that performs temporal graph construction, creates low-dimensional node embeddings, and trains neural network models in an online setting.
arXiv Detail & Related papers (2022-10-08T22:14:31Z) - De Bruijn goes Neural: Causality-Aware Graph Neural Networks for Time
Series Data on Dynamic Graphs [1.2891210250935143]
We introduce De Bruijn Graph Neural Networks (DBGNNs) for time-resolved data on dynamic graphs.
Our approach accounts for temporal-topological patterns that unfold in the causal topology of dynamic graphs.
DBGNNs can leverage temporal patterns in dynamic graphs, which substantially improves the performance in a supervised node classification task.
arXiv Detail & Related papers (2022-09-17T10:54:00Z) - Self-Supervised Dynamic Graph Representation Learning via Temporal
Subgraph Contrast [0.8379286663107846]
This paper proposes a self-supervised dynamic graph representation learning framework (DySubC)
DySubC defines a temporal subgraph contrastive learning task to simultaneously learn the structural and evolutional features of a dynamic graph.
Experiments on five real-world datasets demonstrate that DySubC performs better than the related baselines.
arXiv Detail & Related papers (2021-12-16T09:35:34Z) - Dynamic Graph Learning-Neural Network for Multivariate Time Series
Modeling [2.3022070933226217]
We propose a novel framework, namely static- and dynamic-graph learning-neural network (GL)
The model acquires static and dynamic graph matrices from data to model long-term and short-term patterns respectively.
It achieves state-of-the-art performance on almost all datasets.
arXiv Detail & Related papers (2021-12-06T08:19:15Z) - Temporal Graph Network Embedding with Causal Anonymous Walks
Representations [54.05212871508062]
We propose a novel approach for dynamic network representation learning based on Temporal Graph Network.
For evaluation, we provide a benchmark pipeline for the evaluation of temporal network embeddings.
We show the applicability and superior performance of our model in the real-world downstream graph machine learning task provided by one of the top European banks.
arXiv Detail & Related papers (2021-08-19T15:39:52Z) - Multilevel Graph Matching Networks for Deep Graph Similarity Learning [79.3213351477689]
We propose a multi-level graph matching network (MGMN) framework for computing the graph similarity between any pair of graph-structured objects.
To compensate for the lack of standard benchmark datasets, we have created and collected a set of datasets for both the graph-graph classification and graph-graph regression tasks.
Comprehensive experiments demonstrate that MGMN consistently outperforms state-of-the-art baseline models on both the graph-graph classification and graph-graph regression tasks.
arXiv Detail & Related papers (2020-07-08T19:48:19Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z) - Structural Temporal Graph Neural Networks for Anomaly Detection in
Dynamic Graphs [54.13919050090926]
We propose an end-to-end structural temporal Graph Neural Network model for detecting anomalous edges in dynamic graphs.
In particular, we first extract the $h$-hop enclosing subgraph centered on the target edge and propose the node labeling function to identify the role of each node in the subgraph.
Based on the extracted features, we utilize Gated recurrent units (GRUs) to capture the temporal information for anomaly detection.
arXiv Detail & Related papers (2020-05-15T09:17:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.