Community detection using fast low-cardinality semidefinite programming
- URL: http://arxiv.org/abs/2012.02676v1
- Date: Fri, 4 Dec 2020 15:46:30 GMT
- Title: Community detection using fast low-cardinality semidefinite programming
- Authors: Po-Wei Wang, J. Zico Kolter
- Abstract summary: We propose a new low-cardinality algorithm that generalizes the local update to maximize a semidefinite relaxation derived from Leiden-k-cut.
This proposed algorithm is scalable, outperforms state-of-the-art algorithms, and outperforms in real-world time with little additional cost.
- Score: 94.4878715085334
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modularity maximization has been a fundamental tool for understanding the
community structure of a network, but the underlying optimization problem is
nonconvex and NP-hard to solve. State-of-the-art algorithms like the Louvain or
Leiden methods focus on different heuristics to help escape local optima, but
they still depend on a greedy step that moves node assignment locally and is
prone to getting trapped. In this paper, we propose a new class of
low-cardinality algorithm that generalizes the local update to maximize a
semidefinite relaxation derived from max-k-cut. This proposed algorithm is
scalable, empirically achieves the global semidefinite optimality for small
cases, and outperforms the state-of-the-art algorithms in real-world datasets
with little additional time cost. From the algorithmic perspective, it also
opens a new avenue for scaling-up semidefinite programming when the solutions
are sparse instead of low-rank.
Related papers
- Characterization of Locality in Spin States and Forced Moves for
Optimizations [0.36868085124383626]
In optimization problems, the existence of local minima in energy landscapes is problematic to use to seek the global minimum.
We develop an algorithm to get out of the local minima efficiently while it does not yield the exact samplings.
As the proposed algorithm is based on a rejection-free algorithm, the computational cost is low.
arXiv Detail & Related papers (2023-12-05T07:21:00Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - DESTRESS: Computation-Optimal and Communication-Efficient Decentralized
Nonconvex Finite-Sum Optimization [43.31016937305845]
Internet-of-things, networked sensing, autonomous systems and federated learning call for decentralized algorithms for finite-sum optimizations.
We develop DEcentralized STochastic REcurSive methodDESTRESS for non finite-sum optimization.
Detailed theoretical and numerical comparisons show that DESTRESS improves upon prior decentralized algorithms.
arXiv Detail & Related papers (2021-10-04T03:17:41Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - FedPD: A Federated Learning Framework with Optimal Rates and Adaptivity
to Non-IID Data [59.50904660420082]
Federated Learning (FL) has become a popular paradigm for learning from distributed data.
To effectively utilize data at different devices without moving them to the cloud, algorithms such as the Federated Averaging (FedAvg) have adopted a "computation then aggregation" (CTA) model.
arXiv Detail & Related papers (2020-05-22T23:07:42Z) - Second-Order Guarantees in Centralized, Federated and Decentralized
Nonconvex Optimization [64.26238893241322]
Simple algorithms have been shown to lead to good empirical results in many contexts.
Several works have pursued rigorous analytical justification for studying non optimization problems.
A key insight in these analyses is that perturbations play a critical role in allowing local descent algorithms.
arXiv Detail & Related papers (2020-03-31T16:54:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.