On the inadequacy of nominal assortativity for assessing homophily in
networks
- URL: http://arxiv.org/abs/2211.10245v2
- Date: Tue, 5 Sep 2023 14:12:40 GMT
- Title: On the inadequacy of nominal assortativity for assessing homophily in
networks
- Authors: Fariba Karimi and Marcos Oliveira
- Abstract summary: We show that nominal assortativity fails to account for group imbalance and asymmetric group interactions.
We propose adjusted nominal assortativity and show that this adjustment recovers the expected assortativity in networks with various level of mixing.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nominal assortativity (or discrete assortativity) is widely used to
characterize group mixing patterns and homophily in networks, enabling
researchers to analyze how groups interact with one another. Here we
demonstrate that the measure presents severe shortcomings when applied to
networks with unequal group sizes and asymmetric mixing. We characterize these
shortcomings analytically and use synthetic and empirical networks to show that
nominal assortativity fails to account for group imbalance and asymmetric group
interactions, thereby producing an inaccurate characterization of mixing
patterns. We propose adjusted nominal assortativity and show that this
adjustment recovers the expected assortativity in networks with various level
of mixing. Furthermore, we propose an analytical method to assess asymmetric
mixing by estimating the tendency of inter- and intra-group connectivities.
Finally, we discuss how this approach enables uncovering hidden mixing patterns
in real-world networks.
Related papers
- Dynamic Post-Hoc Neural Ensemblers [55.15643209328513]
In this study, we explore employing neural networks as ensemble methods.
Motivated by the risk of learning low-diversity ensembles, we propose regularizing the model by randomly dropping base model predictions.
We demonstrate this approach lower bounds the diversity within the ensemble, reducing overfitting and improving generalization capabilities.
arXiv Detail & Related papers (2024-10-06T15:25:39Z) - HeNCler: Node Clustering in Heterophilous Graphs through Learned Asymmetric Similarity [55.27586970082595]
HeNCler is a novel approach for Heterophilous Node Clustering.
We show that HeNCler significantly enhances performance in node clustering tasks within heterophilous graph contexts.
arXiv Detail & Related papers (2024-05-27T11:04:05Z) - Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - Pushing Boundaries: Mixup's Influence on Neural Collapse [3.6919724596215615]
Mixup is a data augmentation strategy that employs convex combinations of training instances and their respective labels to augment the robustness and calibration of deep neural networks.
This study investigates the last-layer activations of training data for deep networks subjected to mixup.
We show that mixup's last-layer activations predominantly converge to a distinctive configuration different than one might expect.
arXiv Detail & Related papers (2024-02-09T04:01:25Z) - Consistency for constrained maximum likelihood estimation and clustering based on mixtures of elliptically-symmetric distributions under general data generating processes [0.0]
In a situation where $P$ is a mixture of well enough separated but nonparametric distributions it is shown that the components of the population version of the estimator correspond to the well separated components of $P$.
This provides some theoretical justification for the use of such estimators for cluster analysis in case that $P$ has well separated subpopulations even if these subpopulations differ from what the mixture model assumes.
arXiv Detail & Related papers (2023-11-10T15:20:39Z) - Signed Network Embedding with Application to Simultaneous Detection of
Communities and Anomalies [25.541992448747695]
This paper develops a unified embedding model for signed networks to disentangle the intertwined balance structure and anomaly effect.
The proposed model captures both balance structure and anomaly effect through a low rank plus sparse matrix decomposition.
The advantage of the proposed embedding model is also demonstrated through extensive numerical experiments on both synthetic networks and an international relation network.
arXiv Detail & Related papers (2022-07-08T03:58:56Z) - NOTMAD: Estimating Bayesian Networks with Sample-Specific Structures and
Parameters [70.55488722439239]
We present NOTMAD, which learns to mix archetypal networks according to sample context.
We demonstrate the utility of NOTMAD and sample-specific network inference through analysis and experiments, including patient-specific gene expression networks.
arXiv Detail & Related papers (2021-11-01T17:17:34Z) - Mixture Representation Learning with Coupled Autoencoders [1.589915930948668]
We propose an unsupervised variational framework using multiple interacting networks called cpl-mixVAE.
In this framework, the mixture representation of each network is regularized by imposing a consensus constraint on the discrete factor.
We use the proposed method to jointly uncover discrete and continuous factors of variability describing gene expression in a single-cell transcriptomic dataset.
arXiv Detail & Related papers (2020-07-20T04:12:04Z) - Consistent Estimation of Identifiable Nonparametric Mixture Models from
Grouped Observations [84.81435917024983]
This work proposes an algorithm that consistently estimates any identifiable mixture model from grouped observations.
A practical implementation is provided for paired observations, and the approach is shown to outperform existing methods.
arXiv Detail & Related papers (2020-06-12T20:44:22Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.