GTrans: Spatiotemporal Autoregressive Transformer with Graph Embeddings
for Nowcasting Extreme Events
- URL: http://arxiv.org/abs/2201.06717v1
- Date: Tue, 18 Jan 2022 03:26:24 GMT
- Title: GTrans: Spatiotemporal Autoregressive Transformer with Graph Embeddings
for Nowcasting Extreme Events
- Authors: Bo Feng and Geoffrey Fox
- Abstract summary: This paper proposes atemporal model, namely GTrans, that transforms data features into graph embeddings and predicts temporal dynamics with a transformer model.
According to our experiments, we demonstrate that GTrans can model spatial and temporal dynamics and nowcasts extreme events for datasets.
- Score: 5.672898304129217
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spatiotemporal time series nowcasting should preserve temporal and spatial
dynamics in the sense that generated new sequences from models respect the
covariance relationship from history. Conventional feature extractors are built
with deep convolutional neural networks (CNN). However, CNN models have limits
to image-like applications where data can be formed with high-dimensional
arrays. In contrast, applications in social networks, road traffic, physics,
and chemical property prediction where data features can be organized with
nodes and edges of graphs. Transformer architecture is an emerging method for
predictive models, bringing high accuracy and efficiency due to attention
mechanism design. This paper proposes a spatiotemporal model, namely GTrans,
that transforms data features into graph embeddings and predicts temporal
dynamics with a transformer model. According to our experiments, we demonstrate
that GTrans can model spatial and temporal dynamics and nowcasts extreme events
for datasets. Furthermore, in all the experiments, GTrans can achieve the
highest F1 and F2 scores in binary-class prediction tests than the baseline
models.
Related papers
- DyG-Mamba: Continuous State Space Modeling on Dynamic Graphs [59.434893231950205]
Dynamic graph learning aims to uncover evolutionary laws in real-world systems.
We propose DyG-Mamba, a new continuous state space model for dynamic graph learning.
We show that DyG-Mamba achieves state-of-the-art performance on most datasets.
arXiv Detail & Related papers (2024-08-13T15:21:46Z) - Interpretable A-posteriori Error Indication for Graph Neural Network Surrogate Models [0.0]
This work introduces an interpretability enhancement procedure for graph neural networks (GNNs)
The end result is an interpretable GNN model that isolates regions in physical space, corresponding to sub-graphs, that are intrinsically linked to the forecasting task.
The interpretable GNNs can also be used to identify, during inference, graph nodes that correspond to a majority of the anticipated forecasting error.
arXiv Detail & Related papers (2023-11-13T18:37:07Z) - FourierGNN: Rethinking Multivariate Time Series Forecasting from a Pure
Graph Perspective [48.00240550685946]
Current state-of-the-art graph neural network (GNN)-based forecasting methods usually require both graph networks (e.g., GCN) and temporal networks (e.g., LSTM) to capture inter-series (spatial) dynamics and intra-series (temporal) dependencies, respectively.
We propose a novel Fourier Graph Neural Network (FourierGNN) by stacking our proposed Fourier Graph Operator (FGO) to perform matrix multiplications in Fourier space.
Our experiments on seven datasets have demonstrated superior performance with higher efficiency and fewer parameters compared with state-of-the-
arXiv Detail & Related papers (2023-11-10T17:13:26Z) - Dynamic Causal Explanation Based Diffusion-Variational Graph Neural
Network for Spatio-temporal Forecasting [60.03169701753824]
We propose a novel Dynamic Diffusion-al Graph Neural Network (DVGNN) fortemporal forecasting.
The proposed DVGNN model outperforms state-of-the-art approaches and achieves outstanding Root Mean Squared Error result.
arXiv Detail & Related papers (2023-05-16T11:38:19Z) - A Multidimensional Graph Fourier Transformation Neural Network for
Vehicle Trajectory Prediction [9.554569082679151]
This work introduces the multidimensional Graph Fourier Transformation Neural Network (GFTNN) for long-term trajectory predictions on highways.
Similar to Graph Neural Networks (GNNs), the GFTNN is a novel architecture that operates on graph structures.
For experiments and evaluation, the publicly available datasets highD and NGSIM are used.
arXiv Detail & Related papers (2023-05-12T12:36:48Z) - Dynamic Molecular Graph-based Implementation for Biophysical Properties
Prediction [9.112532782451233]
We propose a novel approach based on the transformer model utilizing GNNs for characterizing dynamic features of protein-ligand interactions.
Our message passing transformer pre-trains on a set of molecular dynamic data based off of physics-based simulations to learn coordinate construction and make binding probability and affinity predictions.
arXiv Detail & Related papers (2022-12-20T04:21:19Z) - An advanced spatio-temporal convolutional recurrent neural network for
storm surge predictions [73.4962254843935]
We study the capability of artificial neural network models to emulate storm surge based on the storm track/size/intensity history.
This study presents a neural network model that can predict storm surge, informed by a database of synthetic storm simulations.
arXiv Detail & Related papers (2022-04-18T23:42:18Z) - PGCN: Progressive Graph Convolutional Networks for Spatial-Temporal Traffic Forecasting [4.14360329494344]
We propose a novel traffic forecasting framework called Progressive Graph Convolutional Network (PGCN)
PGCN constructs a set of graphs by progressively adapting to online input data during the training and testing phases.
The proposed model achieves state-of-the-art performance with consistency in all datasets.
arXiv Detail & Related papers (2022-02-18T02:15:44Z) - TCL: Transformer-based Dynamic Graph Modelling via Contrastive Learning [87.38675639186405]
We propose a novel graph neural network approach, called TCL, which deals with the dynamically-evolving graph in a continuous-time fashion.
To the best of our knowledge, this is the first attempt to apply contrastive learning to representation learning on dynamic graphs.
arXiv Detail & Related papers (2021-05-17T15:33:25Z) - Hyperbolic Variational Graph Neural Network for Modeling Dynamic Graphs [77.33781731432163]
We learn dynamic graph representation in hyperbolic space, for the first time, which aims to infer node representations.
We present a novel Hyperbolic Variational Graph Network, referred to as HVGNN.
In particular, to model the dynamics, we introduce a Temporal GNN (TGNN) based on a theoretically grounded time encoding approach.
arXiv Detail & Related papers (2021-04-06T01:44:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.