Dynamic Spatiotemporal Graph Neural Network with Tensor Network
- URL: http://arxiv.org/abs/2003.08729v1
- Date: Thu, 12 Mar 2020 20:47:22 GMT
- Title: Dynamic Spatiotemporal Graph Neural Network with Tensor Network
- Authors: Chengcheng Jia, Bo Wu, Xiao-Ping Zhang
- Abstract summary: Dynamic spatial graph construction is a challenge in graph neural network (GNN) for time series data problems.
We generate a spatial tensor graph (STG) to collect all the dynamic spatial relations, as well as a temporal tensor graph (TTG) to find the latent pattern along time at each node.
We experimentally compare the accuracy and time costing with the state-of-the-art GNN based methods on the public traffic datasets.
- Score: 12.278768477060137
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic spatial graph construction is a challenge in graph neural network
(GNN) for time series data problems. Although some adaptive graphs are
conceivable, only a 2D graph is embedded in the network to reflect the current
spatial relation, regardless of all the previous situations. In this work, we
generate a spatial tensor graph (STG) to collect all the dynamic spatial
relations, as well as a temporal tensor graph (TTG) to find the latent pattern
along time at each node. These two tensor graphs share the same nodes and
edges, which leading us to explore their entangled correlations by Projected
Entangled Pair States (PEPS) to optimize the two graphs. We experimentally
compare the accuracy and time costing with the state-of-the-art GNN based
methods on the public traffic datasets.
Related papers
- Dynamic Causal Explanation Based Diffusion-Variational Graph Neural
Network for Spatio-temporal Forecasting [60.03169701753824]
We propose a novel Dynamic Diffusion-al Graph Neural Network (DVGNN) fortemporal forecasting.
The proposed DVGNN model outperforms state-of-the-art approaches and achieves outstanding Root Mean Squared Error result.
arXiv Detail & Related papers (2023-05-16T11:38:19Z) - Temporal Aggregation and Propagation Graph Neural Networks for Dynamic
Representation [67.26422477327179]
Temporal graphs exhibit dynamic interactions between nodes over continuous time.
We propose a novel method of temporal graph convolution with the whole neighborhood.
Our proposed TAP-GNN outperforms existing temporal graph methods by a large margin in terms of both predictive performance and online inference latency.
arXiv Detail & Related papers (2023-04-15T08:17:18Z) - Graph Transformer GANs for Graph-Constrained House Generation [223.739067413952]
We present a novel graph Transformer generative adversarial network (GTGAN) to learn effective graph node relations.
The GTGAN learns effective graph node relations in an end-to-end fashion for the challenging graph-constrained house generation task.
arXiv Detail & Related papers (2023-03-14T20:35:45Z) - Multi-Task Edge Prediction in Temporally-Dynamic Video Graphs [16.121140184388786]
We propose MTD-GNN, a graph network for predicting temporally-dynamic edges for multiple types of relations.
We show that modeling multiple relations in our temporal-dynamic graph network can be mutually beneficial.
arXiv Detail & Related papers (2022-12-06T10:41:00Z) - Spatial-Temporal Adaptive Graph Convolution with Attention Network for
Traffic Forecasting [4.1700160312787125]
We propose a novel network, Spatial-Temporal Adaptive graph convolution with Attention Network (STAAN) for traffic forecasting.
Firstly, we adopt an adaptive dependency matrix instead of using a pre-defined matrix during GCN processing to infer the inter-dependencies among nodes.
Secondly, we integrate PW-attention based on graph attention network which is designed for global dependency, and GCN as spatial block.
arXiv Detail & Related papers (2022-06-07T09:08:35Z) - Continuous Temporal Graph Networks for Event-Based Graph Data [41.786721257905555]
We propose Continuous Temporal Graph Networks (CTGNs) to capture the continuous dynamics of temporal graph data.
Key idea is to use neural ordinary differential equations (ODE) to characterize the continuous dynamics of node representations over dynamic graphs.
Experiment results on both transductive and inductive tasks demonstrate the effectiveness of our proposed approach.
arXiv Detail & Related papers (2022-05-31T16:17:02Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - Spatio-Temporal Latent Graph Structure Learning for Traffic Forecasting [6.428566223253948]
We propose a new traffic forecasting framework--S-Temporal Latent Graph Structure Learning networks (ST-LGSL)
The model employs a graph based on Multilayer perceptron and K-Nearest Neighbor, which learns the latent graph topological information from the entire data.
With the dependencies-kNN based on ground-truth adjacency matrix and similarity metric in kNN, ST-LGSL aggregates the top focusing on geography and node similarity.
arXiv Detail & Related papers (2022-02-25T10:02:49Z) - Spatio-Temporal Joint Graph Convolutional Networks for Traffic
Forecasting [75.10017445699532]
Recent have shifted their focus towards formulating traffic forecasting as atemporal graph modeling problem.
We propose a novel approach for accurate traffic forecasting on road networks over multiple future time steps.
arXiv Detail & Related papers (2021-11-25T08:45:14Z) - Structural Temporal Graph Neural Networks for Anomaly Detection in
Dynamic Graphs [54.13919050090926]
We propose an end-to-end structural temporal Graph Neural Network model for detecting anomalous edges in dynamic graphs.
In particular, we first extract the $h$-hop enclosing subgraph centered on the target edge and propose the node labeling function to identify the role of each node in the subgraph.
Based on the extracted features, we utilize Gated recurrent units (GRUs) to capture the temporal information for anomaly detection.
arXiv Detail & Related papers (2020-05-15T09:17:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.