Artificial Benchmark for Community Detection with Outliers (ABCD+o)
- URL: http://arxiv.org/abs/2301.05749v2
- Date: Mon, 12 Jun 2023 18:43:35 GMT
- Title: Artificial Benchmark for Community Detection with Outliers (ABCD+o)
- Authors: Bogumi{\l} Kami\'nski, Pawe{\l} Pra{\l}at, Fran\c{c}ois Th\'eberge
- Abstract summary: We extend the ABCD model to include potential outliers.
We perform some exploratory experiments on both the new ABCD+o model as well as a real-world network to show that outliers possess some desired, distinguishable properties.
- Score: 5.8010446129208155
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Artificial Benchmark for Community Detection graph (ABCD) is a random
graph model with community structure and power-law distribution for both
degrees and community sizes. The model generates graphs with similar properties
as the well-known LFR one, and its main parameter $\xi$ can be tuned to mimic
its counterpart in the LFR model, the mixing parameter $\mu$. In this paper, we
extend the ABCD model to include potential outliers. We perform some
exploratory experiments on both the new ABCD+o model as well as a real-world
network to show that outliers possess some desired, distinguishable properties.
Related papers
- Self-similarity of Communities of the ABCD Model [0.5461938536945723]
The Artificial Benchmark for Community Detection (ABCD) graph is a random graph model with community structure and power-law distribution for both degrees and community sizes.
We show that the ABCD model exhibits some interesting self-similar behaviour.
arXiv Detail & Related papers (2023-11-30T22:52:39Z) - Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution [67.9215891673174]
We propose score entropy as a novel loss that naturally extends score matching to discrete spaces.
We test our Score Entropy Discrete Diffusion models on standard language modeling tasks.
arXiv Detail & Related papers (2023-10-25T17:59:12Z) - Hypergraph Artificial Benchmark for Community Detection (h-ABCD) [5.8010446129208155]
h-ABCD produces random hypergraphs with distributions of ground-truth community sizes and degrees following power-law.
As in the original ABCD, the new model h-ABCD can produce hypergraphs with various levels of noise.
arXiv Detail & Related papers (2022-10-26T20:06:56Z) - Learning from aggregated data with a maximum entropy model [73.63512438583375]
We show how a new model, similar to a logistic regression, may be learned from aggregated data only by approximating the unobserved feature distribution with a maximum entropy hypothesis.
We present empirical evidence on several public datasets that the model learned this way can achieve performances comparable to those of a logistic model trained with the full unaggregated data.
arXiv Detail & Related papers (2022-10-05T09:17:27Z) - uGLAD: Sparse graph recovery by optimizing deep unrolled networks [11.48281545083889]
We present a novel technique to perform sparse graph recovery by optimizing deep unrolled networks.
Our model, uGLAD, builds upon and extends the state-of-the-art model GLAD to the unsupervised setting.
We evaluate model results on synthetic Gaussian data, non-Gaussian data generated from Gene Regulatory Networks, and present a case study in anaerobic digestion.
arXiv Detail & Related papers (2022-05-23T20:20:27Z) - Properties and Performance of the ABCDe Random Graph Model with
Community Structure [4.724825031148412]
We propose a new implementation of the ABCD graph generator, ABCDe, that uses multiple-threading.
We show that ABCDe is more than ten times faster and scales better than the parallel implementation of LFR provided in NetworKit.
arXiv Detail & Related papers (2022-03-28T17:03:24Z) - Modularity of the ABCD Random Graph Model with Community Structure [2.580765958706854]
The ABCD graph is a random graph model with community structure and power-law distribution for both degrees and community sizes.
We analyze the modularity function, arguably the most important graph property of networks in the context of community detection.
It is also used as a quality function in many community detection algorithms, including the widely used Louvain algorithm.
arXiv Detail & Related papers (2022-03-03T01:49:46Z) - Bayesian Structure Learning with Generative Flow Networks [85.84396514570373]
In Bayesian structure learning, we are interested in inferring a distribution over the directed acyclic graph (DAG) from data.
Recently, a class of probabilistic models, called Generative Flow Networks (GFlowNets), have been introduced as a general framework for generative modeling.
We show that our approach, called DAG-GFlowNet, provides an accurate approximation of the posterior over DAGs.
arXiv Detail & Related papers (2022-02-28T15:53:10Z) - Robust Finite Mixture Regression for Heterogeneous Targets [70.19798470463378]
We propose an FMR model that finds sample clusters and jointly models multiple incomplete mixed-type targets simultaneously.
We provide non-asymptotic oracle performance bounds for our model under a high-dimensional learning framework.
The results show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-10-12T03:27:07Z) - Block-Approximated Exponential Random Graphs [77.4792558024487]
An important challenge in the field of exponential random graphs (ERGs) is the fitting of non-trivial ERGs on large graphs.
We propose an approximative framework to such non-trivial ERGs that result in dyadic independence (i.e., edge independent) distributions.
Our methods are scalable to sparse graphs consisting of millions of nodes.
arXiv Detail & Related papers (2020-02-14T11:42:16Z) - Particle-Gibbs Sampling For Bayesian Feature Allocation Models [77.57285768500225]
Most widely used MCMC strategies rely on an element wise Gibbs update of the feature allocation matrix.
We have developed a Gibbs sampler that can update an entire row of the feature allocation matrix in a single move.
This sampler is impractical for models with a large number of features as the computational complexity scales exponentially in the number of features.
We develop a Particle Gibbs sampler that targets the same distribution as the row wise Gibbs updates, but has computational complexity that only grows linearly in the number of features.
arXiv Detail & Related papers (2020-01-25T22:11:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.