Knowledge Generation -- Variational Bayes on Knowledge Graphs
- URL: http://arxiv.org/abs/2101.08857v1
- Date: Thu, 21 Jan 2021 21:23:17 GMT
- Title: Knowledge Generation -- Variational Bayes on Knowledge Graphs
- Authors: Florian Wolf
- Abstract summary: This thesis is a proof of concept for potential of Vari Auto-Encoder (VAE) on representation of real-world Knowledge Graphs.
Inspired by successful approaches to generation graphs, we evaluate the capabilities of our model, the Variational Auto-Encoder (RGVAE)
The RGVAE is first evaluated on link prediction. The mean reciprocal rank (MRR) scores on the two FB15K-237 and WN18RR datasets are compared.
We investigate the latent space in a twofold experiment: first, linear between the latent representation of two triples, then the exploration of each
- Score: 0.685316573653194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This thesis is a proof of concept for the potential of Variational
Auto-Encoder (VAE) on representation learning of real-world Knowledge Graphs
(KG). Inspired by successful approaches to the generation of molecular graphs,
we evaluate the capabilities of our model, the Relational Graph Variational
Auto-Encoder (RGVAE). The impact of the modular hyperparameter choices,
encoding through graph convolutions, graph matching and latent space prior, is
compared. The RGVAE is first evaluated on link prediction. The mean reciprocal
rank (MRR) scores on the two datasets FB15K-237 and WN18RR are compared to the
embedding-based model DistMult. A variational DistMult and a RGVAE without
latent space prior constraint are implemented as control models. The results
show that between different settings, the RGVAE with relaxed latent space,
scores highest on both datasets, yet does not outperform the DistMult. Further,
we investigate the latent space in a twofold experiment: first, linear
interpolation between the latent representation of two triples, then the
exploration of each latent dimension in a $95\%$ confidence interval. Both
interpolations show that the RGVAE learns to reconstruct the adjacency matrix
but fails to disentangle. For the last experiment we introduce a new validation
method for the FB15K-237 data set. The relation type-constrains of generated
triples are filtered and matched with entity types. The observed rate of valid
generated triples is insignificantly higher than the random threshold. All
generated and valid triples are unseen. A comparison between different latent
space priors, using the $\delta$-VAE method, reveals a decoder collapse.
Finally we analyze the limiting factors of our approach compared to molecule
generation and propose solutions for the decoder collapse and successful
representation learning of multi-relational KGs.
Related papers
- Generation is better than Modification: Combating High Class Homophily Variance in Graph Anomaly Detection [51.11833609431406]
Homophily distribution differences between different classes are significantly greater than those in homophilic and heterophilic graphs.
We introduce a new metric called Class Homophily Variance, which quantitatively describes this phenomenon.
To mitigate its impact, we propose a novel GNN model named Homophily Edge Generation Graph Neural Network (HedGe)
arXiv Detail & Related papers (2024-03-15T14:26:53Z) - S^2Former-OR: Single-Stage Bi-Modal Transformer for Scene Graph Generation in OR [50.435592120607815]
Scene graph generation (SGG) of surgical procedures is crucial in enhancing holistically cognitive intelligence in the operating room (OR)
Previous works have primarily relied on multi-stage learning, where the generated semantic scene graphs depend on intermediate processes with pose estimation and object detection.
In this study, we introduce a novel single-stage bi-modal transformer framework for SGG in the OR, termed S2Former-OR.
arXiv Detail & Related papers (2024-02-22T11:40:49Z) - A Contrastive Variational Graph Auto-Encoder for Node Clustering [10.52321770126932]
State-of-the-art clustering methods have numerous challenges.
Existing VGAEs do not account for the discrepancy between the inference and generative models.
Our solution has two mechanisms to control the trade-off between Feature Randomness and Feature Drift.
arXiv Detail & Related papers (2023-12-28T05:07:57Z) - Understanding Augmentation-based Self-Supervised Representation Learning
via RKHS Approximation and Regression [53.15502562048627]
Recent work has built the connection between self-supervised learning and the approximation of the top eigenspace of a graph Laplacian operator.
This work delves into a statistical analysis of augmentation-based pretraining.
arXiv Detail & Related papers (2023-06-01T15:18:55Z) - Deep Graph Stream SVDD: Anomaly Detection in Cyber-Physical Systems [17.373668215331737]
We propose a new approach called deep graph vector data description (SVDD) for anomaly detection.
We first use a transformer to preserve both short and long temporal patterns monitoring data in temporal embeddings.
We cluster these embeddings according to sensor type and utilize them to estimate the change in connectivity between various sensors to construct a new weighted graph.
arXiv Detail & Related papers (2023-02-24T22:14:39Z) - Causal Recurrent Variational Autoencoder for Medical Time Series
Generation [12.82521953179345]
We propose causal recurrent variational autoencoder (CR-VAE), a novel generative model that learns a Granger causal graph from a time series x.
Our model consistently outperforms state-of-the-art time series generative models both qualitatively and quantitatively.
arXiv Detail & Related papers (2023-01-16T19:13:33Z) - Heterogenous Ensemble of Models for Molecular Property Prediction [55.91865861896012]
We propose a method for considering different modalities on molecules.
We ensemble these models with a HuberRegressor.
This yields a winning solution to the 2textsuperscriptnd edition of the OGB Large-Scale Challenge (2022)
arXiv Detail & Related papers (2022-11-20T17:25:26Z) - Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - Optimized latent-code selection for explainable conditional
text-to-image GANs [8.26410341981427]
We present a variety of techniques to take a deep look into the latent space and semantic space of the conditional text-to-image GANs model.
We propose a framework for finding good latent codes by utilizing a linear SVM.
arXiv Detail & Related papers (2022-04-27T03:12:55Z) - T-LoHo: A Bayesian Regularization Model for Structured Sparsity and
Smoothness on Graphs [0.0]
In graph-structured data, structured sparsity and smoothness tend to cluster together.
We propose a new prior for high dimensional parameters with graphical relations.
We use it to detect structured sparsity and smoothness simultaneously.
arXiv Detail & Related papers (2021-07-06T10:10:03Z) - Comprehensive Graph-conditional Similarity Preserving Network for
Unsupervised Cross-modal Hashing [97.44152794234405]
Unsupervised cross-modal hashing (UCMH) has become a hot topic recently.
In this paper, we devise a deep graph-neighbor coherence preserving network (DGCPN)
DGCPN regulates comprehensive similarity preserving losses by exploiting three types of data similarities.
arXiv Detail & Related papers (2020-12-25T07:40:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.