Exploring the Representational Power of Graph Autoencoder
- URL: http://arxiv.org/abs/2106.12005v1
- Date: Tue, 22 Jun 2021 18:23:26 GMT
- Title: Exploring the Representational Power of Graph Autoencoder
- Authors: Maroun Haddad and Mohamed Bouguessa
- Abstract summary: We show that the Degree, the Local Clustering Score, the Betweenness Centrality, the Eigenvector Centrality, and Triangle Count are preserved in the first layer of the graph autoencoder.
We also show that a model with such properties can outperform other models on certain downstream tasks, especially when the preserved features are relevant to the task at hand.
- Score: 1.005130974691351
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: While representation learning has yielded a great success on many graph
learning tasks, there is little understanding behind the structures that are
being captured by these embeddings. For example, we wonder if the topological
features, such as the Triangle Count, the Degree of the node and other
centrality measures are concretely encoded in the embeddings. Furthermore, we
ask if the presence of these structures in the embeddings is necessary for a
better performance on the downstream tasks, such as clustering and
classification. To address these questions, we conduct an extensive empirical
study over three classes of unsupervised graph embedding models and seven
different variants of Graph Autoencoders. Our results show that five
topological features: the Degree, the Local Clustering Score, the Betweenness
Centrality, the Eigenvector Centrality, and Triangle Count are concretely
preserved in the first layer of the graph autoencoder that employs the SUM
aggregation rule, under the condition that the model preserves the second-order
proximity. We supplement further evidence for the presence of these features by
revealing a hierarchy in the distribution of the topological features in the
embeddings of the aforementioned model. We also show that a model with such
properties can outperform other models on certain downstream tasks, especially
when the preserved features are relevant to the task at hand. Finally, we
evaluate the suitability of our findings through a test case study related to
social influence prediction.
Related papers
- Feature graphs for interpretable unsupervised tree ensembles: centrality, interaction, and application in disease subtyping [0.24578723416255746]
Feature selection assumes a pivotal role in enhancing model interpretability.
The accuracy gained from aggregating decision trees comes at the expense of interpretability.
The study introduces novel methods to construct feature graphs from unsupervised random forests.
arXiv Detail & Related papers (2024-04-27T12:47:37Z) - Deep Contrastive Graph Learning with Clustering-Oriented Guidance [61.103996105756394]
Graph Convolutional Network (GCN) has exhibited remarkable potential in improving graph-based clustering.
Models estimate an initial graph beforehand to apply GCN.
Deep Contrastive Graph Learning (DCGL) model is proposed for general data clustering.
arXiv Detail & Related papers (2024-02-25T07:03:37Z) - Rank Collapse Causes Over-Smoothing and Over-Correlation in Graph Neural Networks [3.566568169425391]
We show that with increased depth, node representations become dominated by a low-dimensional subspace that depends on the aggregation function but not on the feature transformations.
For all aggregation functions, the rank of the node representations collapses, resulting in over-smoothing for particular aggregation functions.
arXiv Detail & Related papers (2023-08-31T15:22:31Z) - GrannGAN: Graph annotation generative adversarial networks [72.66289932625742]
We consider the problem of modelling high-dimensional distributions and generating new examples of data with complex relational feature structure coherent with a graph skeleton.
The model we propose tackles the problem of generating the data features constrained by the specific graph structure of each data point by splitting the task into two phases.
In the first it models the distribution of features associated with the nodes of the given graph, in the second it complements the edge features conditionally on the node features.
arXiv Detail & Related papers (2022-12-01T11:49:07Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - TopoDetect: Framework for Topological Features Detection in Graph
Embeddings [1.005130974691351]
TopoDetect is a Python package that allows the user to investigate if important topological features are preserved in the embeddings of graph representation models.
The framework enables the visualization of the embeddings according to the distribution of the topological features among the nodes.
arXiv Detail & Related papers (2021-10-08T14:54:53Z) - Temporal Graph Network Embedding with Causal Anonymous Walks
Representations [54.05212871508062]
We propose a novel approach for dynamic network representation learning based on Temporal Graph Network.
For evaluation, we provide a benchmark pipeline for the evaluation of temporal network embeddings.
We show the applicability and superior performance of our model in the real-world downstream graph machine learning task provided by one of the top European banks.
arXiv Detail & Related papers (2021-08-19T15:39:52Z) - Temporal graph-based approach for behavioural entity classification [0.0]
In this study, a two phased approach for exploiting the potential of graph structures in the cybersecurity domain is presented.
The main idea is to convert a network classification problem into a graph-based behavioural one.
We extract these graph structures that can represent the evolution of both normal and attack entities.
Three clustering techniques are applied to the normal entities in order to aggregate similar behaviours, mitigate the imbalance problem and reduce noisy data.
arXiv Detail & Related papers (2021-05-11T06:13:58Z) - Persistent Homology and Graphs Representation Learning [0.7734726150561088]
We study the topological invariant properties encoded in node graph representational embeddings by utilizing tools available in persistent homology.
Our construction effectively defines a unique persistence-based graph descriptor, on both the graph and node levels.
To demonstrate the effectiveness of the proposed method, we study the topological descriptors induced by DeepWalk, Node2Vec and Diff2Vec.
arXiv Detail & Related papers (2021-02-25T15:26:21Z) - Hierarchical Graph Capsule Network [78.4325268572233]
We propose hierarchical graph capsule network (HGCN) that can jointly learn node embeddings and extract graph hierarchies.
To learn the hierarchical representation, HGCN characterizes the part-whole relationship between lower-level capsules (part) and higher-level capsules (whole)
arXiv Detail & Related papers (2020-12-16T04:13:26Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.