Community Detection with Known, Unknown, or Partially Known Auxiliary
Latent Variables
- URL: http://arxiv.org/abs/2301.04088v1
- Date: Sun, 8 Jan 2023 21:09:03 GMT
- Title: Community Detection with Known, Unknown, or Partially Known Auxiliary
Latent Variables
- Authors: Mohammad Esmaeili and Aria Nosratinia
- Abstract summary: In practice, community membership does not completely explain the dependency between the edges of an observation graph.
We study graphs obeying the block model and censored block model with auxiliary latent variables.
We show that exact recovery is possible by semidefinite programming down to the respective maximum likelihood exact recovery threshold.
- Score: 21.35141858359507
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Empirical observations suggest that in practice, community membership does
not completely explain the dependency between the edges of an observation
graph. The residual dependence of the graph edges are modeled in this paper, to
first order, by auxiliary node latent variables that affect the statistics of
the graph edges but carry no information about the communities of interest. We
then study community detection in graphs obeying the stochastic block model and
censored block model with auxiliary latent variables. We analyze the conditions
for exact recovery when these auxiliary latent variables are unknown,
representing unknown nuisance parameters or model mismatch. We also analyze
exact recovery when these secondary latent variables have been either fully or
partially revealed. Finally, we propose a semidefinite programming algorithm
for recovering the desired labels when the secondary labels are either known or
unknown. We show that exact recovery is possible by semidefinite programming
down to the respective maximum likelihood exact recovery threshold.
Related papers
- Nonparametric Partial Disentanglement via Mechanism Sparsity: Sparse
Actions, Interventions and Sparse Temporal Dependencies [58.179981892921056]
This work introduces a novel principle for disentanglement we call mechanism sparsity regularization.
We propose a representation learning method that induces disentanglement by simultaneously learning the latent factors.
We show that the latent factors can be recovered by regularizing the learned causal graph to be sparse.
arXiv Detail & Related papers (2024-01-10T02:38:21Z) - General Identifiability and Achievability for Causal Representation
Learning [33.80247458590611]
The paper establishes identifiability and achievability results using two hard uncoupled interventions per node in the latent causal graph.
For identifiability, the paper establishes that perfect recovery of the latent causal model and variables is guaranteed under uncoupled interventions.
The analysis, additionally, recovers the identifiability result for two hard coupled interventions, that is when metadata about the pair of environments that have the same node intervened is known.
arXiv Detail & Related papers (2023-10-24T01:47:44Z) - Learning nonparametric latent causal graphs with unknown interventions [18.6470340274888]
We establish conditions under which latent causal graphs are nonparametrically identifiable.
We do not assume the number of hidden variables is known, and we show that at most one unknown intervention per hidden variable is needed.
arXiv Detail & Related papers (2023-06-05T14:06:35Z) - Nonparametric Identifiability of Causal Representations from Unknown
Interventions [63.1354734978244]
We study causal representation learning, the task of inferring latent causal variables and their causal relations from mixtures of the variables.
Our goal is to identify both the ground truth latents and their causal graph up to a set of ambiguities which we show to be irresolvable from interventional data.
arXiv Detail & Related papers (2023-06-01T10:51:58Z) - Posterior Collapse and Latent Variable Non-identifiability [54.842098835445]
We propose a class of latent-identifiable variational autoencoders, deep generative models which enforce identifiability without sacrificing flexibility.
Across synthetic and real datasets, latent-identifiable variational autoencoders outperform existing methods in mitigating posterior collapse and providing meaningful representations of the data.
arXiv Detail & Related papers (2023-01-02T06:16:56Z) - Modeling High-Dimensional Data with Unknown Cut Points: A Fusion
Penalized Logistic Threshold Regression [2.520538806201793]
In traditional logistic regression models, the link function is often assumed to be linear and continuous in predictors.
We consider a threshold model that all continuous features are discretized into ordinal levels, which further determine the binary responses.
We find the lasso model is well suited in the problem of early detection and prediction for chronic disease like diabetes.
arXiv Detail & Related papers (2022-02-17T04:16:40Z) - Correlated Stochastic Block Models: Exact Graph Matching with
Applications to Recovering Communities [2.7920304852537527]
We consider the task of learning latent community structure from multiple correlated networks.
We show how one can exactly recover the latent communities using multiple correlated graphs in parameter regimes.
arXiv Detail & Related papers (2021-07-14T15:27:15Z) - Deconfounded Score Method: Scoring DAGs with Dense Unobserved
Confounding [101.35070661471124]
We show that unobserved confounding leaves a characteristic footprint in the observed data distribution that allows for disentangling spurious and causal effects.
We propose an adjusted score-based causal discovery algorithm that may be implemented with general-purpose solvers and scales to high-dimensional problems.
arXiv Detail & Related papers (2021-03-28T11:07:59Z) - Disentangling Observed Causal Effects from Latent Confounders using
Method of Moments [67.27068846108047]
We provide guarantees on identifiability and learnability under mild assumptions.
We develop efficient algorithms based on coupled tensor decomposition with linear constraints to obtain scalable and guaranteed solutions.
arXiv Detail & Related papers (2021-01-17T07:48:45Z) - The Hidden Uncertainty in a Neural Networks Activations [105.4223982696279]
The distribution of a neural network's latent representations has been successfully used to detect out-of-distribution (OOD) data.
This work investigates whether this distribution correlates with a model's epistemic uncertainty, thus indicating its ability to generalise to novel inputs.
arXiv Detail & Related papers (2020-12-05T17:30:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.