Bayesian Detection of Mesoscale Structures in Pathway Data on Graphs
- URL: http://arxiv.org/abs/2301.11120v1
- Date: Mon, 16 Jan 2023 12:45:33 GMT
- Title: Bayesian Detection of Mesoscale Structures in Pathway Data on Graphs
- Authors: Luka V. Petrovi\'c, Vincenzo Perri
- Abstract summary: mesoscale structures are integral part of the abstraction and analysis of complex systems.
They can represent communities in social or citation networks, roles in corporate interactions, or core-periphery structures in transportation networks.
We derive a Bayesian approach that simultaneously models the optimal partitioning of nodes in groups and the optimal higher-order network dynamics.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mesoscale structures are an integral part of the abstraction and analysis of
complex systems. They reveal a node's function in the network, and facilitate
our understanding of the network dynamics. For example, they can represent
communities in social or citation networks, roles in corporate interactions, or
core-periphery structures in transportation networks. We usually detect
mesoscale structures under the assumption of independence of interactions.
Still, in many cases, the interactions invalidate this assumption by occurring
in a specific order. Such patterns emerge in pathway data; to capture them, we
have to model the dependencies between interactions using higher-order network
models. However, the detection of mesoscale structures in higher-order networks
is still under-researched. In this work, we derive a Bayesian approach that
simultaneously models the optimal partitioning of nodes in groups and the
optimal higher-order network dynamics between the groups. In synthetic data we
demonstrate that our method can recover both standard proximity-based
communities and role-based groupings of nodes. In synthetic and real world data
we show that it can compete with baseline techniques, while additionally
providing interpretable abstractions of network dynamics.
Related papers
- Semantic Loss Functions for Neuro-Symbolic Structured Prediction [74.18322585177832]
We discuss the semantic loss, which injects knowledge about such structure, defined symbolically, into training.
It is agnostic to the arrangement of the symbols, and depends only on the semantics expressed thereby.
It can be combined with both discriminative and generative neural models.
arXiv Detail & Related papers (2024-05-12T22:18:25Z) - Topological Neural Networks: Mitigating the Bottlenecks of Graph Neural
Networks via Higher-Order Interactions [1.994307489466967]
This work starts with a theoretical framework to reveal the impact of network's width, depth, and graph topology on the over-squashing phenomena in message-passing neural networks.
The work drifts towards, higher-order interactions and multi-relational inductive biases via Topological Neural Networks.
Inspired by Graph Attention Networks, two topological attention networks are proposed: Simplicial and Cell Attention Networks.
arXiv Detail & Related papers (2024-02-10T08:26:06Z) - Inferring community structure in attributed hypergraphs using stochastic
block models [3.335932527835653]
We develop a statistical framework that incorporates node attribute data into the learning of community structure in a hypergraph.
We demonstrate that our model, which we refer to as HyperNEO, enhances the learning of community structure in synthetic and empirical hypergraphs.
We expect that our framework will broaden the investigation and understanding of higher-order community structure in real-world complex systems.
arXiv Detail & Related papers (2024-01-01T07:31:32Z) - On Neural Architecture Inductive Biases for Relational Tasks [76.18938462270503]
We introduce a simple architecture based on similarity-distribution scores which we name Compositional Network generalization (CoRelNet)
We find that simple architectural choices can outperform existing models in out-of-distribution generalizations.
arXiv Detail & Related papers (2022-06-09T16:24:01Z) - The interplay between ranking and communities in networks [0.0]
We present a generative model based on an interplay between community and hierarchical structures.
It assumes that each node has a preference in the interaction mechanism and nodes with the same preference are more likely to interact.
We demonstrate our method on synthetic and real-world data and compare performance with two standard approaches for community detection and ranking extraction.
arXiv Detail & Related papers (2021-12-23T16:10:28Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Anomaly Detection on Attributed Networks via Contrastive Self-Supervised
Learning [50.24174211654775]
We present a novel contrastive self-supervised learning framework for anomaly detection on attributed networks.
Our framework fully exploits the local information from network data by sampling a novel type of contrastive instance pair.
A graph neural network-based contrastive learning model is proposed to learn informative embedding from high-dimensional attributes and local structure.
arXiv Detail & Related papers (2021-02-27T03:17:20Z) - Learning the Implicit Semantic Representation on Graph-Structured Data [57.670106959061634]
Existing representation learning methods in graph convolutional networks are mainly designed by describing the neighborhood of each node as a perceptual whole.
We propose a Semantic Graph Convolutional Networks (SGCN) that explores the implicit semantics by learning latent semantic-paths in graphs.
arXiv Detail & Related papers (2021-01-16T16:18:43Z) - Layer-stacked Attention for Heterogeneous Network Embedding [0.0]
Layer-stacked ATTention Embedding (LATTE) is an architecture that automatically decomposes higher-order meta relations at each layer.
LATTE offers a more interpretable aggregation scheme for nodes of different types at different neighborhood ranges.
In both transductive and inductive node classification tasks, LATTE can achieve state-of-the-art performance compared to existing approaches.
arXiv Detail & Related papers (2020-09-17T05:13:41Z) - On the use of local structural properties for improving the efficiency
of hierarchical community detection methods [77.34726150561087]
We study how local structural network properties can be used as proxies to improve the efficiency of hierarchical community detection.
We also check the performance impact of network prunings as an ancillary tactic to make hierarchical community detection more efficient.
arXiv Detail & Related papers (2020-09-15T00:16:12Z) - Shift Aggregate Extract Networks [3.3263205689999453]
We introduce an architecture based on deep hierarchical decompositions to learn effective representations of large graphs.
Our framework extends classic R-decompositions used in kernel methods, enabling nested part-of-part relations.
We show empirically that our approach is able to outperform current state-of-the-art graph classification methods on large social network datasets.
arXiv Detail & Related papers (2017-03-16T09:52:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.