Deep Temporal Graph Clustering: A Comprehensive Benchmark and Datasets
- URL: http://arxiv.org/abs/2601.12903v1
- Date: Mon, 19 Jan 2026 09:58:10 GMT
- Title: Deep Temporal Graph Clustering: A Comprehensive Benchmark and Datasets
- Authors: Meng Liu, Ke Liang, Siwei Wang, Xingchen Hu, Sihang Zhou, Xinwang Liu,
- Abstract summary: Temporal Graph Clustering (TGC) is a new task with little attention, focusing on node clustering in temporal graphs.<n>To address these challenges, we propose a comprehensive benchmark, called BenchTGC.<n> Specially, we design a BenchTGC Framework to illustrate the paradigm of temporal graph clustering and improve existing clustering techniques to fit temporal graphs.
- Score: 43.58219036982937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Temporal Graph Clustering (TGC) is a new task with little attention, focusing on node clustering in temporal graphs. Compared with existing static graph clustering, it can find the balance between time requirement and space requirement (Time-Space Balance) through the interaction sequence-based batch-processing pattern. However, there are two major challenges that hinder the development of TGC, i.e., inapplicable clustering techniques and inapplicable datasets. To address these challenges, we propose a comprehensive benchmark, called BenchTGC. Specially, we design a BenchTGC Framework to illustrate the paradigm of temporal graph clustering and improve existing clustering techniques to fit temporal graphs. In addition, we also discuss problems with public temporal graph datasets and develop multiple datasets suitable for TGC task, called BenchTGC Datasets. According to extensive experiments, we not only verify the advantages of BenchTGC, but also demonstrate the necessity and importance of TGC task. We wish to point out that the dynamically changing and complex scenarios in real world are the foundation of temporal graph clustering. The code and data is available at: https://github.com/MGitHubL/BenchTGC.
Related papers
- Federated Temporal Graph Clustering [9.779760673367663]
Temporal graph clustering is a complex task that involves discovering meaningful structures in dynamic graphs where relationships and entities change over time.<n>Existing methods typically require centralized data collection, which poses significant privacy and communication challenges.<n>We introduce a novel Federated Temporal Graph Clustering framework that enables decentralized training of graph neural networks (GNNs) across multiple clients.
arXiv Detail & Related papers (2024-10-16T08:04:57Z) - GC-Bench: An Open and Unified Benchmark for Graph Condensation [54.70801435138878]
We develop a comprehensive Graph Condensation Benchmark (GC-Bench) to analyze the performance of graph condensation.
GC-Bench systematically investigates the characteristics of graph condensation in terms of the following dimensions: effectiveness, transferability, and complexity.
We have developed an easy-to-use library for training and evaluating different GC methods to facilitate reproducible research.
arXiv Detail & Related papers (2024-06-30T07:47:34Z) - CueGCL: Cluster-aware Personalized Self-Training for Unsupervised Graph Contrastive Learning [49.88192702588169]
We propose a Cluster-aware Graph Contrastive Learning Framework (CueGCL) to jointly learn clustering results and node representations.<n>Specifically, we design a personalized self-training (PeST) strategy for unsupervised scenarios, which enables our model to capture precise cluster-level personalized information.<n>We theoretically demonstrate the effectiveness of our model, showing it yields an embedding space with a significantly discernible cluster structure.
arXiv Detail & Related papers (2023-11-18T13:45:21Z) - Temporal Graph Benchmark for Machine Learning on Temporal Graphs [54.52243310226456]
Temporal Graph Benchmark (TGB) is a collection of challenging and diverse benchmark datasets.
We benchmark each dataset and find that the performance of common models can vary drastically across datasets.
TGB provides an automated machine learning pipeline for reproducible and accessible temporal graph research.
arXiv Detail & Related papers (2023-07-03T13:58:20Z) - arXiv4TGC: Large-Scale Datasets for Temporal Graph Clustering [52.63652741011945]
We build arXiv4TGC, a set of novel academic datasets for temporal graph clustering.
In particular, the largest dataset, arXivLarge, contains 1.3 million labeled available nodes and 10 million temporal edges.
The clustering performance on arXiv4TGC can be more apparent for evaluating different models.
arXiv Detail & Related papers (2023-06-08T06:37:04Z) - Deep Temporal Graph Clustering [77.02070768950145]
We propose a general framework for deep Temporal Graph Clustering (GC)
GC introduces deep clustering techniques to suit the interaction sequence-based batch-processing pattern of temporal graphs.
Our framework can effectively improve the performance of existing temporal graph learning methods.
arXiv Detail & Related papers (2023-05-18T06:17:50Z) - Clustering of Time-Varying Graphs Based on Temporal Label Smoothness [28.025212175496964]
We propose a node clustering method for time-varying graphs based on the assumption that the cluster labels are changed smoothly over time.
Experiments on synthetic and real-world time-varying graphs are performed to validate the effectiveness of the proposed approach.
arXiv Detail & Related papers (2023-05-11T05:20:41Z) - GLCC: A General Framework for Graph-level Clustering [5.069852282550117]
This paper studies the problem of graph-level clustering, which is a novel yet challenging task.
We propose a general graph-level clustering framework named Graph-Level Contrastive Clustering (GLCC)
Experiments on a range of well-known datasets demonstrate the superiority of our proposed GLCC over competitive baselines.
arXiv Detail & Related papers (2022-10-21T11:08:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.