Perfect Spectral Clustering with Discrete Covariates
- URL: http://arxiv.org/abs/2205.08047v1
- Date: Tue, 17 May 2022 01:41:06 GMT
- Title: Perfect Spectral Clustering with Discrete Covariates
- Authors: Jonathan Hehir, Xiaoyue Niu, Aleksandra Slavkovic
- Abstract summary: We propose a spectral algorithm that achieves perfect clustering with high probability on a class of large, sparse networks.
Our method is the first to offer a guarantee of consistent latent structure recovery using spectral clustering.
- Score: 68.8204255655161
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Among community detection methods, spectral clustering enjoys two desirable
properties: computational efficiency and theoretical guarantees of consistency.
Most studies of spectral clustering consider only the edges of a network as
input to the algorithm. Here we consider the problem of performing community
detection in the presence of discrete node covariates, where network structure
is determined by a combination of a latent block model structure and homophily
on the observed covariates. We propose a spectral algorithm that we prove
achieves perfect clustering with high probability on a class of large, sparse
networks with discrete covariates, effectively separating latent network
structure from homophily on observed covariates. To our knowledge, our method
is the first to offer a guarantee of consistent latent structure recovery using
spectral clustering in the setting where edge formation is dependent on both
latent and observed factors.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - HeNCler: Node Clustering in Heterophilous Graphs through Learned Asymmetric Similarity [55.27586970082595]
HeNCler is a novel approach for Heterophilous Node Clustering.
We show that HeNCler significantly enhances performance in node clustering tasks within heterophilous graph contexts.
arXiv Detail & Related papers (2024-05-27T11:04:05Z) - Deep Embedding Clustering Driven by Sample Stability [16.53706617383543]
We propose a deep embedding clustering algorithm driven by sample stability (DECS)
Specifically, we start by constructing the initial feature space with an autoencoder and then learn the cluster-oriented embedding feature constrained by sample stability.
The experimental results on five datasets illustrate that the proposed method achieves superior performance compared to state-of-the-art clustering approaches.
arXiv Detail & Related papers (2024-01-29T09:19:49Z) - flow-based clustering and spectral clustering: a comparison [0.688204255655161]
We study a novel graph clustering method for data with an intrinsic network structure.
We exploit an intrinsic network structure of data to construct Euclidean feature vectors.
Our results indicate that our clustering methods can cope with certain graph structures.
arXiv Detail & Related papers (2022-06-20T21:49:52Z) - Spectral clustering via adaptive layer aggregation for multi-layer
networks [6.0073653636512585]
We propose integrative spectral clustering approaches based on effective convex layer aggregations.
We show that our methods are remarkably competitive compared to several popularly used methods.
arXiv Detail & Related papers (2020-12-07T21:58:18Z) - An improved spectral clustering method for community detection under the
degree-corrected stochastic blockmodel [1.0965065178451106]
We propose an improved spectral clustering (ISC) approach under the degree corrected block model (SBM)
ISC provides a significant improvement on two weak signal networks Simmons and Caltech, with error rates of 121/1137 and 96/590, respectively.
arXiv Detail & Related papers (2020-11-12T13:35:11Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Stable and consistent density-based clustering via multiparameter
persistence [77.34726150561087]
We consider the degree-Rips construction from topological data analysis.
We analyze its stability to perturbations of the input data using the correspondence-interleaving distance.
We integrate these methods into a pipeline for density-based clustering, which we call Persistable.
arXiv Detail & Related papers (2020-05-18T19:45:04Z) - Local Graph Clustering with Network Lasso [90.66817876491052]
We study the statistical and computational properties of a network Lasso method for local graph clustering.
The clusters delivered by nLasso can be characterized elegantly via network flows between cluster boundary and seed nodes.
arXiv Detail & Related papers (2020-04-25T17:52:05Z) - Randomized spectral co-clustering for large-scale directed networks [15.486507430729052]
Co-clustering aims to cluster the senders and receivers of directed networks simultaneously.
We leverage sketching techniques and derive two randomized spectral co-clustering algorithms.
We establish their approximation error rates and misclustering error rates, indicating better bounds than the state-of-the-art results of co-clustering literature.
arXiv Detail & Related papers (2020-04-25T15:00:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.