KCoreMotif: An Efficient Graph Clustering Algorithm for Large Networks
by Exploiting k-core Decomposition and Motifs
- URL: http://arxiv.org/abs/2008.10380v1
- Date: Fri, 21 Aug 2020 12:21:05 GMT
- Title: KCoreMotif: An Efficient Graph Clustering Algorithm for Large Networks
by Exploiting k-core Decomposition and Motifs
- Authors: Gang Mei, Jingzhi Tu, Lei Xiao, Francesco Piccialli
- Abstract summary: Spectral clustering is one of the most commonly used algorithms for graph-structured data (networks)
We propose an efficient graph clustering algorithm, KCoreMotif, specifically for large networks by exploiting k-core decomposition and motifs.
Comparative results demonstrate that the proposed graph clustering algorithm is accurate yet efficient for large networks.
- Score: 6.1734015475373765
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Clustering analysis has been widely used in trust evaluation on various
complex networks such as wireless sensors networks and online social networks.
Spectral clustering is one of the most commonly used algorithms for
graph-structured data (networks). However, the conventional spectral clustering
is inherently difficult to work with large-scale networks due to the fact that
it needs computationally expensive matrix manipulations. To deal with large
networks, in this paper, we propose an efficient graph clustering algorithm,
KCoreMotif, specifically for large networks by exploiting k-core decomposition
and motifs. The essential idea behind the proposed clustering algorithm is to
perform the efficient motif-based spectral clustering algorithm on k-core
subgraphs, rather than on the entire graph. More specifically, (1) we first
conduct the k-core decomposition of the large input network; (2) we then
perform the motif-based spectral clustering for the top k-core subgraphs; (3)
we group the remaining vertices in the rest (k-1)-core subgraphs into
previously found clusters; and finally obtain the desired clusters of the large
input network. To evaluate the performance of the proposed graph clustering
algorithm KCoreMotif, we use both the conventional and the motif-based spectral
clustering algorithms as the baselines and compare our algorithm with them for
18 groups of real-world datasets. Comparative results demonstrate that the
proposed graph clustering algorithm is accurate yet efficient for large
networks, which also means that it can be further used to evaluate the
intra-cluster and inter-cluster trusts on large networks.
Related papers
- A Versatile Framework for Attributed Network Clustering via K-Nearest Neighbor Augmentation [14.327262299413789]
We develop ANCKA as a versatile attributed network clustering framework, capable of attributed graph clustering (AGC), attributed multiplex graph clustering (AMGC), and attributed hypergraph clustering (AHC)
We have conducted extensive experiments to compare our methods with 19 competitors on 8 attributed hypergraphs, 16 competitors on 6 attributed graphs, and 16 competitors on 3 attributed multiplex graphs, all demonstrating the superb clustering quality and efficiency of our methods.
arXiv Detail & Related papers (2024-08-10T06:59:51Z) - Deep Contrastive Graph Learning with Clustering-Oriented Guidance [61.103996105756394]
Graph Convolutional Network (GCN) has exhibited remarkable potential in improving graph-based clustering.
Models estimate an initial graph beforehand to apply GCN.
Deep Contrastive Graph Learning (DCGL) model is proposed for general data clustering.
arXiv Detail & Related papers (2024-02-25T07:03:37Z) - Reinforcement Graph Clustering with Unknown Cluster Number [91.4861135742095]
We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
arXiv Detail & Related papers (2023-08-13T18:12:28Z) - DeepCluE: Enhanced Image Clustering via Multi-layer Ensembles in Deep
Neural Networks [53.88811980967342]
This paper presents a Deep Clustering via Ensembles (DeepCluE) approach.
It bridges the gap between deep clustering and ensemble clustering by harnessing the power of multiple layers in deep neural networks.
Experimental results on six image datasets confirm the advantages of DeepCluE over the state-of-the-art deep clustering approaches.
arXiv Detail & Related papers (2022-06-01T09:51:38Z) - A Modular Framework for Centrality and Clustering in Complex Networks [0.6423239719448168]
In this paper, we study two important such network analysis techniques, namely, centrality and clustering.
An information-flow based model is adopted for clustering, which itself builds upon an information theoretic measure for computing centrality.
Our clustering naturally inherits the flexibility to accommodate edge directionality, as well as different interpretations and interplay between edge weights and node degrees.
arXiv Detail & Related papers (2021-11-23T03:01:29Z) - Self-supervised Contrastive Attributed Graph Clustering [110.52694943592974]
We propose a novel attributed graph clustering network, namely Self-supervised Contrastive Attributed Graph Clustering (SCAGC)
In SCAGC, by leveraging inaccurate clustering labels, a self-supervised contrastive loss, are designed for node representation learning.
For the OOS nodes, SCAGC can directly calculate their clustering labels.
arXiv Detail & Related papers (2021-10-15T03:25:28Z) - Graph Neural Networks with Composite Kernels [60.81504431653264]
We re-interpret node aggregation from the perspective of kernel weighting.
We present a framework to consider feature similarity in an aggregation scheme.
We propose feature aggregation as the composition of the original neighbor-based kernel and a learnable kernel to encode feature similarities in a feature space.
arXiv Detail & Related papers (2020-05-16T04:44:29Z) - Hierarchical clustering of bipartite data sets based on the statistical
significance of coincidences [0.0]
We provide an hierarchical clustering algorithm based on a dissimilarity between entities that quantifies the probability that the features shared by two entities is due to mere chance.
The algorithm performance is $O(n2)$ when applied to a set of n entities, and its outcome is a dendrogram exhibiting the connections of those entities.
arXiv Detail & Related papers (2020-04-27T23:30:18Z) - Randomized spectral co-clustering for large-scale directed networks [15.486507430729052]
Co-clustering aims to cluster the senders and receivers of directed networks simultaneously.
We leverage sketching techniques and derive two randomized spectral co-clustering algorithms.
We establish their approximation error rates and misclustering error rates, indicating better bounds than the state-of-the-art results of co-clustering literature.
arXiv Detail & Related papers (2020-04-25T15:00:55Z) - Motif-Based Spectral Clustering of Weighted Directed Networks [6.1448102196124195]
Clustering is an essential technique for network analysis, with applications in a diverse range of fields.
One approach is to capture and cluster higher-order structures using motif adjacency matrices.
We present new and computationally useful matrix formulae for motif adjacency matrices on weighted networks.
arXiv Detail & Related papers (2020-04-02T22:45:28Z) - Learning to Cluster Faces via Confidence and Connectivity Estimation [136.5291151775236]
We propose a fully learnable clustering framework without requiring a large number of overlapped subgraphs.
Our method significantly improves clustering accuracy and thus performance of the recognition models trained on top, yet it is an order of magnitude more efficient than existing supervised methods.
arXiv Detail & Related papers (2020-04-01T13:39:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.