DyTed: Disentangled Representation Learning for Discrete-time Dynamic
Graph
- URL: http://arxiv.org/abs/2210.10592v2
- Date: Wed, 16 Aug 2023 01:01:50 GMT
- Title: DyTed: Disentangled Representation Learning for Discrete-time Dynamic
Graph
- Authors: Kaike Zhang, Qi Cao, Gaolin Fang, Bingbing Xu, Hongjian Zou, Huawei
Shen, Xueqi Cheng
- Abstract summary: We propose a novel disenTangled representation learning framework for discrete-time Dynamic graphs, namely DyTed.
We specially design a temporal-clips contrastive learning task together with a structure contrastive learning to effectively identify the time-invariant and time-varying representations respectively.
- Score: 59.583555454424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised representation learning for dynamic graphs has attracted a lot
of research attention in recent years. Compared with static graph, the dynamic
graph is a comprehensive embodiment of both the intrinsic stable
characteristics of nodes and the time-related dynamic preference. However,
existing methods generally mix these two types of information into a single
representation space, which may lead to poor explanation, less robustness, and
a limited ability when applied to different downstream tasks. To solve the
above problems, in this paper, we propose a novel disenTangled representation
learning framework for discrete-time Dynamic graphs, namely DyTed. We specially
design a temporal-clips contrastive learning task together with a structure
contrastive learning to effectively identify the time-invariant and
time-varying representations respectively. To further enhance the
disentanglement of these two types of representation, we propose a
disentanglement-aware discriminator under an adversarial learning framework
from the perspective of information theory. Extensive experiments on Tencent
and five commonly used public datasets demonstrate that DyTed, as a general
framework that can be applied to existing methods, achieves state-of-the-art
performance on various downstream tasks, as well as be more robust against
noise.
Related papers
- TimeGraphs: Graph-based Temporal Reasoning [64.18083371645956]
TimeGraphs is a novel approach that characterizes dynamic interactions as a hierarchical temporal graph.
Our approach models the interactions using a compact graph-based representation, enabling adaptive reasoning across diverse time scales.
We evaluate TimeGraphs on multiple datasets with complex, dynamic agent interactions, including a football simulator, the Resistance game, and the MOMA human activity dataset.
arXiv Detail & Related papers (2024-01-06T06:26:49Z) - Temporal Graph Representation Learning with Adaptive Augmentation
Contrastive [12.18909612212823]
Temporal graph representation learning aims to generate low-dimensional dynamic node embeddings to capture temporal information.
We propose a novel Temporal Graph representation learning with Adaptive augmentation Contrastive (TGAC) model.
Our experiments on various real networks demonstrate that the proposed model outperforms other temporal graph representation learning methods.
arXiv Detail & Related papers (2023-11-07T11:21:16Z) - From random-walks to graph-sprints: a low-latency node embedding
framework on continuous-time dynamic graphs [4.372841335228306]
We propose a framework for continuous-time-dynamic-graphs (CTDGs) that has low latency and is competitive with state-of-the-art, higher latency models.
In our framework, time-aware node embeddings summarizing multi-hop information are computed using only single-hop operations on the incoming edges.
We demonstrate that our graph-sprints features, combined with a machine learning, achieve competitive performance.
arXiv Detail & Related papers (2023-07-17T12:25:52Z) - Intensity Profile Projection: A Framework for Continuous-Time
Representation Learning for Dynamic Networks [50.2033914945157]
We present a representation learning framework, Intensity Profile Projection, for continuous-time dynamic network data.
The framework consists of three stages: estimating pairwise intensity functions, learning a projection which minimises a notion of intensity reconstruction error.
Moreoever, we develop estimation theory providing tight control on the error of any estimated trajectory, indicating that the representations could even be used in quite noise-sensitive follow-on analyses.
arXiv Detail & Related papers (2023-06-09T15:38:25Z) - Self-Supervised Temporal Graph learning with Temporal and Structural Intensity Alignment [53.72873672076391]
Temporal graph learning aims to generate high-quality representations for graph-based tasks with dynamic information.
We propose a self-supervised method called S2T for temporal graph learning, which extracts both temporal and structural information.
S2T achieves at most 10.13% performance improvement compared with the state-of-the-art competitors on several datasets.
arXiv Detail & Related papers (2023-02-15T06:36:04Z) - Deep Dynamic Effective Connectivity Estimation from Multivariate Time
Series [0.0]
We develop dynamic effective connectivity estimation via neural network training (DECENNT)
DECENNT outperforms state-of-the-art (SOTA) methods on five different tasks and infers interpretable task-specific dynamic graphs.
arXiv Detail & Related papers (2022-02-04T21:14:21Z) - Self-Supervised Dynamic Graph Representation Learning via Temporal
Subgraph Contrast [0.8379286663107846]
This paper proposes a self-supervised dynamic graph representation learning framework (DySubC)
DySubC defines a temporal subgraph contrastive learning task to simultaneously learn the structural and evolutional features of a dynamic graph.
Experiments on five real-world datasets demonstrate that DySubC performs better than the related baselines.
arXiv Detail & Related papers (2021-12-16T09:35:34Z) - Group Contrastive Self-Supervised Learning on Graphs [101.45974132613293]
We study self-supervised learning on graphs using contrastive methods.
We argue that contrasting graphs in multiple subspaces enables graph encoders to capture more abundant characteristics.
arXiv Detail & Related papers (2021-07-20T22:09:21Z) - TCL: Transformer-based Dynamic Graph Modelling via Contrastive Learning [87.38675639186405]
We propose a novel graph neural network approach, called TCL, which deals with the dynamically-evolving graph in a continuous-time fashion.
To the best of our knowledge, this is the first attempt to apply contrastive learning to representation learning on dynamic graphs.
arXiv Detail & Related papers (2021-05-17T15:33:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.