Reinforcement Graph Clustering with Unknown Cluster Number
- URL: http://arxiv.org/abs/2308.06827v1
- Date: Sun, 13 Aug 2023 18:12:28 GMT
- Title: Reinforcement Graph Clustering with Unknown Cluster Number
- Authors: Yue Liu, Ke Liang, Jun Xia, Xihong Yang, Sihang Zhou, Meng Liu,
Xinwang Liu, Stan Z. Li
- Abstract summary: We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
- Score: 91.4861135742095
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep graph clustering, which aims to group nodes into disjoint clusters by
neural networks in an unsupervised manner, has attracted great attention in
recent years. Although the performance has been largely improved, the excellent
performance of the existing methods heavily relies on an accurately predefined
cluster number, which is not always available in the real-world scenario. To
enable the deep graph clustering algorithms to work without the guidance of the
predefined cluster number, we propose a new deep graph clustering method termed
Reinforcement Graph Clustering (RGC). In our proposed method, cluster number
determination and unsupervised representation learning are unified into a
uniform framework by the reinforcement learning mechanism. Concretely, the
discriminative node representations are first learned with the contrastive
pretext task. Then, to capture the clustering state accurately with both local
and global information in the graph, both node and cluster states are
considered. Subsequently, at each state, the qualities of different cluster
numbers are evaluated by the quality network, and the greedy action is executed
to determine the cluster number. In order to conduct feedback actions, the
clustering-oriented reward function is proposed to enhance the cohesion of the
same clusters and separate the different clusters. Extensive experiments
demonstrate the effectiveness and efficiency of our proposed method. The source
code of RGC is shared at https://github.com/yueliu1999/RGC and a collection
(papers, codes and, datasets) of deep graph clustering is shared at
https://github.com/yueliu1999/Awesome-Deep-Graph-Clustering on Github.
Related papers
- DGCLUSTER: A Neural Framework for Attributed Graph Clustering via
Modularity Maximization [5.329981192545312]
We propose a novel method, DGCluster, which primarily optimize the modularity objective using graph neural networks and scales linearly with the graph size.
We extensively test DGCluster on several real-world datasets of varying sizes, across multiple popular cluster quality metrics.
Our approach consistently outperforms the state-of-the-art methods, demonstrating significant performance gains in almost all settings.
arXiv Detail & Related papers (2023-12-20T01:43:55Z) - Learning Uniform Clusters on Hypersphere for Deep Graph-level Clustering [25.350054742471816]
We propose a novel deep graph-level clustering method called Uniform Deep Graph Clustering (UDGC)
UDGC assigns instances evenly to different clusters and then scatters those clusters on unit hypersphere, leading to a more uniform cluster-level distribution and a slighter cluster collapse.
Our empirical study on eight well-known datasets demonstrates that UDGC significantly outperforms the state-of-the-art models.
arXiv Detail & Related papers (2023-11-23T12:08:20Z) - Dink-Net: Neural Clustering on Large Graphs [59.10189693120368]
A deep graph clustering method (Dink-Net) is proposed with the idea of dilation and shrink.
By discriminating nodes, whether being corrupted by augmentations, representations are learned in a self-supervised manner.
The clustering distribution is optimized by minimizing the proposed cluster dilation loss and cluster shrink loss.
Compared to the runner-up, Dink-Net 9.62% achieves NMI improvement on the ogbn-papers100M dataset with 111 million nodes and 1.6 billion edges.
arXiv Detail & Related papers (2023-05-28T15:33:24Z) - GLCC: A General Framework for Graph-level Clustering [5.069852282550117]
This paper studies the problem of graph-level clustering, which is a novel yet challenging task.
We propose a general graph-level clustering framework named Graph-Level Contrastive Clustering (GLCC)
Experiments on a range of well-known datasets demonstrate the superiority of our proposed GLCC over competitive baselines.
arXiv Detail & Related papers (2022-10-21T11:08:10Z) - DeepCluE: Enhanced Image Clustering via Multi-layer Ensembles in Deep
Neural Networks [53.88811980967342]
This paper presents a Deep Clustering via Ensembles (DeepCluE) approach.
It bridges the gap between deep clustering and ensemble clustering by harnessing the power of multiple layers in deep neural networks.
Experimental results on six image datasets confirm the advantages of DeepCluE over the state-of-the-art deep clustering approaches.
arXiv Detail & Related papers (2022-06-01T09:51:38Z) - Graph Representation Learning via Contrasting Cluster Assignments [57.87743170674533]
We propose a novel unsupervised graph representation model by contrasting cluster assignments, called as GRCCA.
It is motivated to make good use of local and global information synthetically through combining clustering algorithms and contrastive learning.
GRCCA has strong competitiveness in most tasks.
arXiv Detail & Related papers (2021-12-15T07:28:58Z) - Self-supervised Contrastive Attributed Graph Clustering [110.52694943592974]
We propose a novel attributed graph clustering network, namely Self-supervised Contrastive Attributed Graph Clustering (SCAGC)
In SCAGC, by leveraging inaccurate clustering labels, a self-supervised contrastive loss, are designed for node representation learning.
For the OOS nodes, SCAGC can directly calculate their clustering labels.
arXiv Detail & Related papers (2021-10-15T03:25:28Z) - Structured Graph Learning for Clustering and Semi-supervised
Classification [74.35376212789132]
We propose a graph learning framework to preserve both the local and global structure of data.
Our method uses the self-expressiveness of samples to capture the global structure and adaptive neighbor approach to respect the local structure.
Our model is equivalent to a combination of kernel k-means and k-means methods under certain condition.
arXiv Detail & Related papers (2020-08-31T08:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.