From random-walks to graph-sprints: a low-latency node embedding
framework on continuous-time dynamic graphs
- URL: http://arxiv.org/abs/2307.08433v5
- Date: Fri, 16 Feb 2024 23:34:24 GMT
- Title: From random-walks to graph-sprints: a low-latency node embedding
framework on continuous-time dynamic graphs
- Authors: Ahmad Naser Eddin, Jacopo Bono, David Apar\'icio, Hugo Ferreira,
Jo\~ao Ascens\~ao, Pedro Ribeiro, Pedro Bizarro
- Abstract summary: We propose a framework for continuous-time-dynamic-graphs (CTDGs) that has low latency and is competitive with state-of-the-art, higher latency models.
In our framework, time-aware node embeddings summarizing multi-hop information are computed using only single-hop operations on the incoming edges.
We demonstrate that our graph-sprints features, combined with a machine learning, achieve competitive performance.
- Score: 4.372841335228306
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Many real-world datasets have an underlying dynamic graph structure, where
entities and their interactions evolve over time. Machine learning models
should consider these dynamics in order to harness their full potential in
downstream tasks. Previous approaches for graph representation learning have
focused on either sampling k-hop neighborhoods, akin to breadth-first search,
or random walks, akin to depth-first search. However, these methods are
computationally expensive and unsuitable for real-time, low-latency inference
on dynamic graphs. To overcome these limitations, we propose graph-sprints a
general purpose feature extraction framework for continuous-time-dynamic-graphs
(CTDGs) that has low latency and is competitive with state-of-the-art, higher
latency models. To achieve this, a streaming, low latency approximation to the
random-walk based features is proposed. In our framework, time-aware node
embeddings summarizing multi-hop information are computed using only single-hop
operations on the incoming edges. We evaluate our proposed approach on three
open-source datasets and two in-house datasets, and compare with three
state-of-the-art algorithms (TGN-attn, TGN-ID, Jodie). We demonstrate that our
graph-sprints features, combined with a machine learning classifier, achieve
competitive performance (outperforming all baselines for the node
classification tasks in five datasets). Simultaneously, graph-sprints
significantly reduce inference latencies, achieving close to an order of
magnitude speed-up in our experimental setting.
Related papers
- TempoKGAT: A Novel Graph Attention Network Approach for Temporal Graph Analysis [3.5707423185282656]
This paper presents a new type of graph attention network, called TempoKGAT, which combines time-decaying weight and a selective neighbor aggregation mechanism on the spatial domain.
We evaluate our approach on multiple datasets from the traffic, energy, and health sectors involvingtemporal data.
arXiv Detail & Related papers (2024-08-29T09:54:46Z) - Node-Time Conditional Prompt Learning In Dynamic Graphs [14.62182210205324]
We propose DYGPROMPT, a novel pre-training and prompt learning framework for dynamic graph modeling.
We recognize that node and time features mutually characterize each other, and propose dual condition-nets to model the evolving node-time patterns in downstream tasks.
arXiv Detail & Related papers (2024-05-22T19:10:24Z) - Input Snapshots Fusion for Scalable Discrete Dynamic Graph Nerual Networks [27.616083395612595]
We introduce an Input bf Snapshots bf Fusion based bf Dynamic bf Graph Neural Network (SFDyG)
By eliminating the partitioning of snapshots within the input window, we obtain a multi-graph (more than one edge between two nodes)
We propose a scalable three-step mini-batch training method and demonstrate its equivalence to full-batch training counterpart.
arXiv Detail & Related papers (2024-05-11T10:05:55Z) - TimeGraphs: Graph-based Temporal Reasoning [64.18083371645956]
TimeGraphs is a novel approach that characterizes dynamic interactions as a hierarchical temporal graph.
Our approach models the interactions using a compact graph-based representation, enabling adaptive reasoning across diverse time scales.
We evaluate TimeGraphs on multiple datasets with complex, dynamic agent interactions, including a football simulator, the Resistance game, and the MOMA human activity dataset.
arXiv Detail & Related papers (2024-01-06T06:26:49Z) - SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - Sparsity exploitation via discovering graphical models in multi-variate
time-series forecasting [1.2762298148425795]
We propose a decoupled training method, which includes a graph generating module and a GNNs forecasting module.
First, we use Graphical Lasso (or GraphLASSO) to directly exploit the sparsity pattern from data to build graph structures.
Second, we fit these graph structures and the input data into a Graph Convolutional Recurrent Network (GCRN) to train a forecasting model.
arXiv Detail & Related papers (2023-06-29T16:48:00Z) - Deep Temporal Graph Clustering [77.02070768950145]
We propose a general framework for deep Temporal Graph Clustering (GC)
GC introduces deep clustering techniques to suit the interaction sequence-based batch-processing pattern of temporal graphs.
Our framework can effectively improve the performance of existing temporal graph learning methods.
arXiv Detail & Related papers (2023-05-18T06:17:50Z) - Temporal Aggregation and Propagation Graph Neural Networks for Dynamic
Representation [67.26422477327179]
Temporal graphs exhibit dynamic interactions between nodes over continuous time.
We propose a novel method of temporal graph convolution with the whole neighborhood.
Our proposed TAP-GNN outperforms existing temporal graph methods by a large margin in terms of both predictive performance and online inference latency.
arXiv Detail & Related papers (2023-04-15T08:17:18Z) - Multivariate Time Series Forecasting with Dynamic Graph Neural ODEs [65.18780403244178]
We propose a continuous model to forecast Multivariate Time series with dynamic Graph neural Ordinary Differential Equations (MTGODE)
Specifically, we first abstract multivariate time series into dynamic graphs with time-evolving node features and unknown graph structures.
Then, we design and solve a neural ODE to complement missing graph topologies and unify both spatial and temporal message passing.
arXiv Detail & Related papers (2022-02-17T02:17:31Z) - Dynamic Graph Learning-Neural Network for Multivariate Time Series
Modeling [2.3022070933226217]
We propose a novel framework, namely static- and dynamic-graph learning-neural network (GL)
The model acquires static and dynamic graph matrices from data to model long-term and short-term patterns respectively.
It achieves state-of-the-art performance on almost all datasets.
arXiv Detail & Related papers (2021-12-06T08:19:15Z) - Spatio-Temporal Joint Graph Convolutional Networks for Traffic
Forecasting [75.10017445699532]
Recent have shifted their focus towards formulating traffic forecasting as atemporal graph modeling problem.
We propose a novel approach for accurate traffic forecasting on road networks over multiple future time steps.
arXiv Detail & Related papers (2021-11-25T08:45:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.