SSTKG: Simple Spatio-Temporal Knowledge Graph for Intepretable and
Versatile Dynamic Information Embedding
- URL: http://arxiv.org/abs/2402.12132v1
- Date: Mon, 19 Feb 2024 13:28:43 GMT
- Title: SSTKG: Simple Spatio-Temporal Knowledge Graph for Intepretable and
Versatile Dynamic Information Embedding
- Authors: Ruiyi Yang, Flora D. Salim and Hao Xue
- Abstract summary: This paper introduces a novel framework for constructing and exploring on temporal-temporal Knowledge (STSKG)
Our framework offers a simple but comprehensive way to understand the underlying trends in dynamic KG thereby enhancing the accuracy of predictions and the relevance of recommendations.
- Score: 11.919765478453964
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Knowledge graphs (KGs) have been increasingly employed for link prediction
and recommendation using real-world datasets. However, the majority of current
methods rely on static data, neglecting the dynamic nature and the hidden
spatio-temporal attributes of real-world scenarios. This often results in
suboptimal predictions and recommendations. Although there are effective
spatio-temporal inference methods, they face challenges such as scalability
with large datasets and inadequate semantic understanding, which impede their
performance. To address these limitations, this paper introduces a novel
framework - Simple Spatio-Temporal Knowledge Graph (SSTKG), for constructing
and exploring spatio-temporal KGs. To integrate spatial and temporal data into
KGs, our framework exploited through a new 3-step embedding method. Output
embeddings can be used for future temporal sequence prediction and spatial
information recommendation, providing valuable insights for various
applications such as retail sales forecasting and traffic volume prediction.
Our framework offers a simple but comprehensive way to understand the
underlying patterns and trends in dynamic KG, thereby enhancing the accuracy of
predictions and the relevance of recommendations. This work paves the way for
more effective utilization of spatio-temporal data in KGs, with potential
impacts across a wide range of sectors.
Related papers
- Conservation-informed Graph Learning for Spatiotemporal Dynamics Prediction [84.26340606752763]
In this paper, we introduce the conservation-informed GNN (CiGNN), an end-to-end explainable learning framework.
The network is designed to conform to the general symmetry conservation law via symmetry where conservative and non-conservative information passes over a multiscale space by a latent temporal marching strategy.
Results demonstrate that CiGNN exhibits remarkable baseline accuracy and generalizability, and is readily applicable to learning for prediction of varioustemporal dynamics.
arXiv Detail & Related papers (2024-12-30T13:55:59Z) - ST-FiT: Inductive Spatial-Temporal Forecasting with Limited Training Data [59.78770412981611]
In real-world applications, most nodes may not possess any available temporal data during training.
We propose a principled framework named ST-FiT to handle this problem.
arXiv Detail & Related papers (2024-12-14T17:51:29Z) - EasyST: A Simple Framework for Spatio-Temporal Prediction [18.291117879544945]
We propose a simple framework for spatial-temporal prediction - EasyST paradigm.
It learns lightweight and robust Multi-Layer Perceptrons (MLPs) generalization by distilling knowledge from complex-temporal GNNs.
EasyST surpasses state-of-the-art approaches in terms of efficiency and accuracy.
arXiv Detail & Related papers (2024-09-10T11:40:01Z) - Retrieval-Augmented Generation Meets Data-Driven Tabula Rasa Approach for Temporal Knowledge Graph Forecasting [0.0]
sLA-tKGF is a small-scale language assistant for temporal Knowledge Graph (tKG) forecasting.
Our framework constructs knowledge-infused prompts with historical data from tKGs and web search results.
It reduces hallucinations and mitigates distributional shift challenges through comprehending changing trends over time.
arXiv Detail & Related papers (2024-08-18T11:52:24Z) - DyG-Mamba: Continuous State Space Modeling on Dynamic Graphs [59.434893231950205]
Dynamic graph learning aims to uncover evolutionary laws in real-world systems.
We propose DyG-Mamba, a new continuous state space model for dynamic graph learning.
We show that DyG-Mamba achieves state-of-the-art performance on most datasets.
arXiv Detail & Related papers (2024-08-13T15:21:46Z) - Physics-guided Active Sample Reweighting for Urban Flow Prediction [75.24539704456791]
Urban flow prediction is a nuanced-temporal modeling that estimates the throughput of transportation services like buses, taxis and ride-driven models.
Some recent prediction solutions bring remedies with the notion of physics-guided machine learning (PGML)
We develop a atized physics-guided network (PN), and propose a data-aware framework Physics-guided Active Sample Reweighting (P-GASR)
arXiv Detail & Related papers (2024-07-18T15:44:23Z) - UrbanGPT: Spatio-Temporal Large Language Models [34.79169613947957]
We present the UrbanPT, which seamlessly integrates atemporal-temporal encoder with instruction-tuning paradigm.
We conduct extensive experiments on various public datasets, covering differenttemporal prediction tasks.
The results consistently demonstrate that our UrbanPT, with its carefully designed architecture, consistently outperforms state-of-the-art baselines.
arXiv Detail & Related papers (2024-02-25T12:37:29Z) - A Simple Temporal Information Matching Mechanism for Entity Alignment
Between Temporal Knowledge Graphs [18.451872649228196]
We propose a simple graph neural network (GNN) model combined with a temporal information matching mechanism.
We also propose a method to generate unsupervised alignment seeds via the temporal information of TKG.
arXiv Detail & Related papers (2022-09-20T12:27:34Z) - SLPC: a VRNN-based approach for stochastic lidar prediction and
completion in autonomous driving [63.87272273293804]
We propose a new LiDAR prediction framework that is based on generative models namely Variational Recurrent Neural Networks (VRNNs)
Our algorithm is able to address the limitations of previous video prediction frameworks when dealing with sparse data by spatially inpainting the depth maps in the upcoming frames.
We present a sparse version of VRNNs and an effective self-supervised training method that does not require any labels.
arXiv Detail & Related papers (2021-02-19T11:56:44Z) - From Static to Dynamic Node Embeddings [61.58641072424504]
We introduce a general framework for leveraging graph stream data for temporal prediction-based applications.
Our proposed framework includes novel methods for learning an appropriate graph time-series representation.
We find that the top-3 temporal models are always those that leverage the new $epsilon$-graph time-series representation.
arXiv Detail & Related papers (2020-09-21T16:48:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.