The Importance of Being Correlated: Implications of Dependence in Joint
Spectral Inference across Multiple Networks
- URL: http://arxiv.org/abs/2008.00163v3
- Date: Thu, 17 Jun 2021 17:10:29 GMT
- Title: The Importance of Being Correlated: Implications of Dependence in Joint
Spectral Inference across Multiple Networks
- Authors: Konstantinos Pantazis, Avanti Athreya, Jes\'us Arroyo, William N.
Frost, Evan S. Hill, and Vince Lyzinski
- Abstract summary: Spectral inference on multiple networks is a rapidly-developing subfield of graph statistics.
Recent work has demonstrated that joint, or simultaneous, spectral embedding of multiple independent networks can deliver more accurate estimation.
We present a generalized omnibus embedding methodology and provide a detailed analysis of this embedding across both independent and correlated networks.
- Score: 4.238478445823
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spectral inference on multiple networks is a rapidly-developing subfield of
graph statistics. Recent work has demonstrated that joint, or simultaneous,
spectral embedding of multiple independent networks can deliver more accurate
estimation than individual spectral decompositions of those same networks. Such
inference procedures typically rely heavily on independence assumptions across
the multiple network realizations, and even in this case, little attention has
been paid to the induced network correlation in such joint embeddings. Here, we
present a generalized omnibus embedding methodology and provide a detailed
analysis of this embedding across both independent and correlated networks, the
latter of which significantly extends the reach of such procedures. We describe
how this omnibus embedding can itself induce correlation, leading us to
distinguish between inherent correlation -- the correlation that arises
naturally in multisample network data -- and induced correlation, which is an
artifice of the joint embedding methodology. We show that the generalized
omnibus embedding procedure is flexible and robust, and prove both consistency
and a central limit theorem for the embedded points. We examine how induced and
inherent correlation can impact inference for network time series data, and we
provide network analogues of classical questions such as the effective sample
size for more generally correlated data. Further, we show how an appropriately
calibrated generalized omnibus embedding can detect changes in real biological
networks that previous embedding procedures could not discern, confirming that
the effect of inherent and induced correlation can be subtle and
transformative, with import in theory and practice.
Related papers
- Network Causal Effect Estimation In Graphical Models Of Contagion And Latent Confounding [2.654975444537834]
Key question in many network studies is whether the observed correlations between units are primarily due to contagion or latent confounding.
We propose network causal effect estimation strategies that provide unbiased and consistent estimates.
We evaluate the effectiveness of our methods with synthetic data and the validity of our assumptions using real-world networks.
arXiv Detail & Related papers (2024-11-02T22:12:44Z) - Intrinsic Dimension Correlation: uncovering nonlinear connections in multimodal representations [0.4223422932643755]
This paper exploits the entanglement between intrinsic dimensionality and correlation to propose a metric that quantifies the correlation.
We first validate our method on synthetic data in controlled environments, showcasing its advantages and drawbacks compared to existing techniques.
We extend our analysis to large-scale applications in neural network representations.
arXiv Detail & Related papers (2024-06-22T10:36:04Z) - Learning Divergence Fields for Shift-Robust Graph Representations [73.11818515795761]
In this work, we propose a geometric diffusion model with learnable divergence fields for the challenging problem with interdependent data.
We derive a new learning objective through causal inference, which can guide the model to learn generalizable patterns of interdependence that are insensitive across domains.
arXiv Detail & Related papers (2024-06-07T14:29:21Z) - Concentration inequalities for correlated network-valued processes with
applications to community estimation and changepoint analysis [2.1953732467962315]
We study the aggregate behavior of network sequences generated from network-valued processes.
We demonstrate the usefulness of these concentration results in proving consistency of standard estimators in community estimation and changepoint estimation problems.
We also conduct a simulation study to demonstrate the effect of the laziness parameter, which controls the extent of temporal correlation, on the accuracy of community and changepoint estimation.
arXiv Detail & Related papers (2022-08-02T11:16:58Z) - On Neural Architecture Inductive Biases for Relational Tasks [76.18938462270503]
We introduce a simple architecture based on similarity-distribution scores which we name Compositional Network generalization (CoRelNet)
We find that simple architectural choices can outperform existing models in out-of-distribution generalizations.
arXiv Detail & Related papers (2022-06-09T16:24:01Z) - Decomposing neural networks as mappings of correlation functions [57.52754806616669]
We study the mapping between probability distributions implemented by a deep feed-forward network.
We identify essential statistics in the data, as well as different information representations that can be used by neural networks.
arXiv Detail & Related papers (2022-02-10T09:30:31Z) - Prequential MDL for Causal Structure Learning with Neural Networks [9.669269791955012]
We show that the prequential minimum description length principle can be used to derive a practical scoring function for Bayesian networks.
We obtain plausible and parsimonious graph structures without relying on sparsity inducing priors or other regularizers which must be tuned.
We discuss how the the prequential score relates to recent work that infers causal structure from the speed of adaptation when the observations come from a source undergoing distributional shift.
arXiv Detail & Related papers (2021-07-02T22:35:21Z) - Decentralized Local Stochastic Extra-Gradient for Variational
Inequalities [125.62877849447729]
We consider distributed variational inequalities (VIs) on domains with the problem data that is heterogeneous (non-IID) and distributed across many devices.
We make a very general assumption on the computational network that covers the settings of fully decentralized calculations.
We theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone settings.
arXiv Detail & Related papers (2021-06-15T17:45:51Z) - Deep Archimedean Copulas [98.96141706464425]
ACNet is a novel differentiable neural network architecture that enforces structural properties.
We show that ACNet is able to both approximate common Archimedean Copulas and generate new copulas which may provide better fits to data.
arXiv Detail & Related papers (2020-12-05T22:58:37Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.