Fast Network Community Detection with Profile-Pseudo Likelihood Methods
- URL: http://arxiv.org/abs/2011.00647v3
- Date: Sun, 29 Aug 2021 15:32:14 GMT
- Title: Fast Network Community Detection with Profile-Pseudo Likelihood Methods
- Authors: Jiangzhou Wang, Jingfei Zhang, Binghui Liu, Ji Zhu, and Jianhua Guo
- Abstract summary: Most algorithms for fitting the block model likelihood function cannot scale to large-scale networks.
We propose a novel likelihood approach that decouples row and column labels in the likelihood function.
We show that our method provides strongly consistent estimates of the communities in a block model.
- Score: 19.639557431997037
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The stochastic block model is one of the most studied network models for
community detection. It is well-known that most algorithms proposed for fitting
the stochastic block model likelihood function cannot scale to large-scale
networks. One prominent work that overcomes this computational challenge is
Amini et al.(2013), which proposed a fast pseudo-likelihood approach for
fitting stochastic block models to large sparse networks. However, this
approach does not have convergence guarantee, and is not well suited for small-
or medium- scale networks. In this article, we propose a novel likelihood based
approach that decouples row and column labels in the likelihood function, which
enables a fast alternating maximization; the new method is computationally
efficient, performs well for both small and large scale networks, and has
provable convergence guarantee. We show that our method provides strongly
consistent estimates of the communities in a stochastic block model. As
demonstrated in simulation studies, the proposed method outperforms the
pseudo-likelihood approach in terms of both estimation accuracy and computation
efficiency, especially for large sparse networks. We further consider
extensions of our proposed method to handle networks with degree heterogeneity
and bipartite properties.
Related papers
- Sampling and active learning methods for network reliability estimation using K-terminal spanning tree [16.985964958558586]
Network reliability analysis remains a challenge due to the increasing size and complexity of networks.
This paper presents a novel sampling method and an active learning method for efficient and accurate network reliability estimation.
arXiv Detail & Related papers (2024-07-09T08:51:53Z) - Fast Shapley Value Estimation: A Unified Approach [71.92014859992263]
We propose a straightforward and efficient Shapley estimator, SimSHAP, by eliminating redundant techniques.
In our analysis of existing approaches, we observe that estimators can be unified as a linear transformation of randomly summed values from feature subsets.
Our experiments validate the effectiveness of our SimSHAP, which significantly accelerates the computation of accurate Shapley values.
arXiv Detail & Related papers (2023-11-02T06:09:24Z) - Implicit Variational Inference for High-Dimensional Posteriors [7.924706533725115]
In variational inference, the benefits of Bayesian models rely on accurately capturing the true posterior distribution.
We propose using neural samplers that specify implicit distributions, which are well-suited for approximating complex multimodal and correlated posteriors.
Our approach introduces novel bounds for approximate inference using implicit distributions by locally linearising the neural sampler.
arXiv Detail & Related papers (2023-10-10T14:06:56Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Distributed Dynamic Safe Screening Algorithms for Sparse Regularization [73.85961005970222]
We propose a new distributed dynamic safe screening (DDSS) method for sparsity regularized models and apply it on shared-memory and distributed-memory architecture respectively.
We prove that the proposed method achieves the linear convergence rate with lower overall complexity and can eliminate almost all the inactive features in a finite number of iterations almost surely.
arXiv Detail & Related papers (2022-04-23T02:45:55Z) - Bayesian community detection for networks with covariates [16.230648949593153]
"Community detection" has arguably received the most attention in the scientific community.
We propose a block model with a co-dependent random partition prior.
Our model has the ability to learn the number of the communities via posterior inference without having to assume it to be known.
arXiv Detail & Related papers (2022-03-04T01:58:35Z) - Network Estimation by Mixing: Adaptivity and More [2.3478438171452014]
We propose a mixing strategy that leverages available arbitrary models to improve their individual performances.
The proposed method is computationally efficient and almost tuning-free.
We show that the proposed method performs equally well as the oracle estimate when the true model is included as individual candidates.
arXiv Detail & Related papers (2021-06-05T05:17:04Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Efficient semidefinite-programming-based inference for binary and
multi-class MRFs [83.09715052229782]
We propose an efficient method for computing the partition function or MAP estimate in a pairwise MRF.
We extend semidefinite relaxations from the typical binary MRF to the full multi-class setting, and develop a compact semidefinite relaxation that can again be solved efficiently using the solver.
arXiv Detail & Related papers (2020-12-04T15:36:29Z) - Learning All Credible Bayesian Network Structures for Model Averaging [3.81379858342235]
We propose a novel approach to model averaging inspired by performance guarantees in approximation algorithms.
Our approach is more efficient and scales to significantly larger Bayesian networks than existing approaches.
arXiv Detail & Related papers (2020-08-27T19:56:27Z) - ESPN: Extremely Sparse Pruned Networks [50.436905934791035]
We show that a simple iterative mask discovery method can achieve state-of-the-art compression of very deep networks.
Our algorithm represents a hybrid approach between single shot network pruning methods and Lottery-Ticket type approaches.
arXiv Detail & Related papers (2020-06-28T23:09:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.