$p$-Norm Flow Diffusion for Local Graph Clustering
- URL: http://arxiv.org/abs/2005.09810v3
- Date: Mon, 20 Jul 2020 09:09:08 GMT
- Title: $p$-Norm Flow Diffusion for Local Graph Clustering
- Authors: Kimon Fountoulakis, Di Wang, Shenghao Yang
- Abstract summary: We present a family of convex optimization formulations based on the idea of diffusion with p-norm network flow for $pin (1,infty)$.
In the context of local clustering, we show their usefulness in finding low conductance cuts around input seed set.
We show that the proposed problem can be solved in strongly local running time for $pge 2$ and conduct empirical evaluations on both synthetic and real-world graphs to illustrate our approach compares favorably with existing methods.
- Score: 18.90840167452118
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Local graph clustering and the closely related seed set expansion problem are
primitives on graphs that are central to a wide range of analytic and learning
tasks such as local clustering, community detection, nodes ranking and feature
inference. Prior work on local graph clustering mostly falls into two
categories with numerical and combinatorial roots respectively. In this work,
we draw inspiration from both fields and propose a family of convex
optimization formulations based on the idea of diffusion with p-norm network
flow for $p\in (1,\infty)$. In the context of local clustering, we characterize
the optimal solutions for these optimization problems and show their usefulness
in finding low conductance cuts around input seed set. In particular, we
achieve quadratic approximation of conductance in the case of $p=2$ similar to
the Cheeger-type bounds of spectral methods, constant factor approximation when
$p\rightarrow\infty$ similar to max-flow based methods, and a smooth transition
for general $p$ values in between. Thus, our optimization formulation can be
viewed as bridging the numerical and combinatorial approaches, and we can
achieve the best of both worlds in terms of speed and noise robustness. We show
that the proposed problem can be solved in strongly local running time for
$p\ge 2$ and conduct empirical evaluations on both synthetic and real-world
graphs to illustrate our approach compares favorably with existing methods.
Related papers
- Towards Efficient Pareto-optimal Utility-Fairness between Groups in
Repeated Rankings [7.6275971668447005]
We tackle the problem of computing a sequence of rankings with the guarantee of the Pareto-optimal balance between consumers and producers of items.
We introduce a novel approach to the above problem by using the Expohedron - a permutahedron whose points represent all exposures of items.
We further propose an efficient method by relaxing our optimization problem on the Expohedron's circumscribed $n$-sphere, which significantly improve the running time.
arXiv Detail & Related papers (2024-02-22T05:48:54Z) - Faster Convergence with Multiway Preferences [99.68922143784306]
We consider the sign-function-based comparison feedback model and analyze the convergence rates with batched and multiway comparisons.
Our work is the first to study the problem of convex optimization with multiway preferences and analyze the optimal convergence rates.
arXiv Detail & Related papers (2023-12-19T01:52:13Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Pick your Neighbor: Local Gauss-Southwell Rule for Fast Asynchronous
Decentralized Optimization [37.85806148420504]
In decentralized optimization environments, each agent $i$ in a network of $n$ optimization nodes possesses a private function $f_i$.
We introduce an optimization-aware selection rule that chooses the neighbor with the highest dual cost improvement.
We show that the proposed set-wise GS rule achieves a speedup by a factor of up to the maximum degree in the network.
arXiv Detail & Related papers (2022-07-15T15:37:03Z) - $\ell_2$-norm Flow Diffusion in Near-Linear Time [18.263667346853698]
We provide an $widetildeO(m)$-time randomized algorithm for the $ell$-norm flow diffusion problem.
It is done simply by sweeping over the dual solution vector.
It adapts the high-level algorithmic framework of Laplacian system solvers, but requires several new tools.
arXiv Detail & Related papers (2021-05-30T21:27:58Z) - Laplace Matching for fast Approximate Inference in Generalized Linear
Models [27.70274403550477]
We propose an approximate inference framework primarily designed to be computationally cheap while still achieving high approximation quality.
The concept, which we call emphLaplace Matching, involves closed-form, approximate, bi-directional transformations between the parameter spaces of exponential families.
This effectively turns inference in GLMs into conjugate inference (with small approximation errors)
arXiv Detail & Related papers (2021-05-07T08:25:17Z) - The Min-Max Complexity of Distributed Stochastic Convex Optimization
with Intermittent Communication [61.60069865891783]
We resolve the min-max complexity of distributed convex optimization (up to a log factor) in the intermittent communication setting.
We present a novel lower bound with a matching upper bound that establishes an optimal algorithm.
arXiv Detail & Related papers (2021-02-02T16:18:29Z) - Local Graph Clustering with Network Lasso [90.66817876491052]
We study the statistical and computational properties of a network Lasso method for local graph clustering.
The clusters delivered by nLasso can be characterized elegantly via network flows between cluster boundary and seed nodes.
arXiv Detail & Related papers (2020-04-25T17:52:05Z) - Stochastic Flows and Geometric Optimization on the Orthogonal Group [52.50121190744979]
We present a new class of geometrically-driven optimization algorithms on the orthogonal group $O(d)$.
We show that our methods can be applied in various fields of machine learning including deep, convolutional and recurrent neural networks, reinforcement learning, flows and metric learning.
arXiv Detail & Related papers (2020-03-30T15:37:50Z) - Optimization of Graph Total Variation via Active-Set-based Combinatorial
Reconditioning [48.42916680063503]
We propose a novel adaptive preconditioning strategy for proximal algorithms on this problem class.
We show that nested-forest decomposition of the inactive edges yields a guaranteed local linear convergence rate.
Our results suggest that local convergence analysis can serve as a guideline for selecting variable metrics in proximal algorithms.
arXiv Detail & Related papers (2020-02-27T16:33:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.