Spatial-temporal Graph Convolutional Networks with Diversified Transformation for Dynamic Graph Representation Learning
- URL: http://arxiv.org/abs/2408.02704v1
- Date: Mon, 5 Aug 2024 09:40:47 GMT
- Title: Spatial-temporal Graph Convolutional Networks with Diversified Transformation for Dynamic Graph Representation Learning
- Authors: Ling Wang, Yixiang Huang, Hao Wu,
- Abstract summary: This study proposes a spatial-temporal graph convolutional networks with diversified transformation (STGCNDT)
It includes three aspects: a) constructing a unified graph tensor convolutional network (GTCN) using tensor M-products without the need to representtemporal information separately; b) introducing three transformation schemes in GTN to model complex temporal patterns to aggregate temporal information; and c) constructing an ensemble of diversified transformations to obtain higher representation capabilities.
- Score: 6.9243139068960895
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic graphs (DG) are often used to describe evolving interactions between nodes in real-world applications. Temporal patterns are a natural feature of DGs and are also key to representation learning. However, existing dynamic GCN models are mostly composed of static GCNs and sequence modules, which results in the separation of spatiotemporal information and cannot effectively capture complex temporal patterns in DGs. To address this problem, this study proposes a spatial-temporal graph convolutional networks with diversified transformation (STGCNDT), which includes three aspects: a) constructing a unified graph tensor convolutional network (GTCN) using tensor M-products without the need to represent spatiotemporal information separately; b) introducing three transformation schemes in GTCN to model complex temporal patterns to aggregate temporal information; and c) constructing an ensemble of diversified transformation schemes to obtain higher representation capabilities. Empirical studies on four DGs that appear in communication networks show that the proposed STGCNDT significantly outperforms state-of-the-art models in solving link weight estimation tasks due to the diversified transformations.
Related papers
- Point Cloud Denoising With Fine-Granularity Dynamic Graph Convolutional Networks [58.050130177241186]
Noise perturbations often corrupt 3-D point clouds, hindering downstream tasks such as surface reconstruction, rendering, and further processing.
This paper introduces finegranularity dynamic graph convolutional networks called GDGCN, a novel approach to denoising in 3-D point clouds.
arXiv Detail & Related papers (2024-11-21T14:19:32Z) - A Novel Spatiotemporal Coupling Graph Convolutional Network [0.18130068086063336]
This paper presents a novel Graph Contemporalal Networks (GCNs)-based dynamic estimator namely Spatio Coupling GCN (SCG) model with the three-fold ideas as below.
The results demonstrate that SCG realizes higher accuracy compared with the state-of-the-arts, illustrating it can learn powerful representations to users and cloud services.
arXiv Detail & Related papers (2024-08-09T02:02:01Z) - DTFormer: A Transformer-Based Method for Discrete-Time Dynamic Graph Representation Learning [38.53424185696828]
The representation learning of Discrete-Time Dynamic Graphs (DTDGs) has been extensively applied to model the dynamics of temporally changing entities and their evolving connections.
This paper introduces a novel representation learning method DTFormer for DTDGs, pivoting from the traditional GNN+RNN framework to a Transformer-based architecture.
arXiv Detail & Related papers (2024-07-26T05:46:23Z) - Learning Invariant Representations of Graph Neural Networks via Cluster
Generalization [58.68231635082891]
Graph neural networks (GNNs) have become increasingly popular in modeling graph-structured data.
In this paper, we experimentally find that the performance of GNNs drops significantly when the structure shift happens.
We propose the Cluster Information Transfer (CIT) mechanism, which can learn invariant representations for GNNs.
arXiv Detail & Related papers (2024-03-06T10:36:56Z) - Tensor Graph Convolutional Network for Dynamic Graph Representation
Learning [12.884025972321316]
Dynamic graphs (DG) describe dynamic interactions between entities in many practical scenarios.
Most existing DG representation learning models combine graph convolutional network and sequence neural network.
We propose a tensor graph convolutional network to learn DG representations in one convolution framework.
arXiv Detail & Related papers (2024-01-13T12:49:56Z) - Dynamic Graph Representation Learning via Edge Temporal States Modeling and Structure-reinforced Transformer [5.093187534912688]
We introduce the Recurrent Structure-reinforced Graph Transformer (RSGT), a novel framework for dynamic graph representation learning.
RSGT captures temporal node representations encoding both graph topology and evolving dynamics through a recurrent learning paradigm.
We show RSGT's superior performance in discrete dynamic graph representation learning, consistently outperforming existing methods in dynamic link prediction tasks.
arXiv Detail & Related papers (2023-04-20T04:12:50Z) - DG-STGCN: Dynamic Spatial-Temporal Modeling for Skeleton-based Action
Recognition [77.87404524458809]
We propose a new framework for skeleton-based action recognition, namely Dynamic Group Spatio-Temporal GCN (DG-STGCN)
It consists of two modules, DG-GCN and DG-TCN, respectively, for spatial and temporal modeling.
DG-STGCN consistently outperforms state-of-the-art methods, often by a notable margin.
arXiv Detail & Related papers (2022-10-12T03:17:37Z) - Multivariate Time Series Forecasting with Dynamic Graph Neural ODEs [65.18780403244178]
We propose a continuous model to forecast Multivariate Time series with dynamic Graph neural Ordinary Differential Equations (MTGODE)
Specifically, we first abstract multivariate time series into dynamic graphs with time-evolving node features and unknown graph structures.
Then, we design and solve a neural ODE to complement missing graph topologies and unify both spatial and temporal message passing.
arXiv Detail & Related papers (2022-02-17T02:17:31Z) - Equivariance-bridged SO(2)-Invariant Representation Learning using Graph
Convolutional Network [0.1657441317977376]
Training a Convolutional Neural Network (CNN) to be robust against rotation has mostly been done with data augmentation.
This paper highlights to encourage less dependence on data augmentation by achieving structural rotational invariance of a network.
Our method achieves the state-of-the-art image classification performance on rotated MNIST and CIFAR-10 images.
arXiv Detail & Related papers (2021-06-18T08:37:45Z) - Self-Supervised Graph Representation Learning via Topology
Transformations [61.870882736758624]
We present the Topology Transformation Equivariant Representation learning, a general paradigm of self-supervised learning for node representations of graph data.
In experiments, we apply the proposed model to the downstream node and graph classification tasks, and results show that the proposed method outperforms the state-of-the-art unsupervised approaches.
arXiv Detail & Related papers (2021-05-25T06:11:03Z) - E(n) Equivariant Graph Neural Networks [86.75170631724548]
This paper introduces a new model to learn graph neural networks equivariant to rotations, translations, reflections and permutations called E(n)-Equivariant Graph Neural Networks (EGNNs)
In contrast with existing methods, our work does not require computationally expensive higher-order representations in intermediate layers while it still achieves competitive or better performance.
arXiv Detail & Related papers (2021-02-19T10:25:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.