Kronecker Decomposition for Knowledge Graph Embeddings
- URL: http://arxiv.org/abs/2205.06560v1
- Date: Fri, 13 May 2022 11:11:03 GMT
- Title: Kronecker Decomposition for Knowledge Graph Embeddings
- Authors: Caglar Demir and Julian Lienen and Axel-Cyrille Ngonga Ngomo
- Abstract summary: We propose a technique based on Kronecker decomposition to reduce the number of parameters in a knowledge graph embedding model.
The decomposition ensures that elementwise interactions between three embedding vectors are extended with interactions within each embedding vector.
Our experiments suggest that applying Kronecker decomposition on embedding matrices leads to an improved parameter efficiency on all benchmark datasets.
- Score: 5.49810117202384
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Knowledge graph embedding research has mainly focused on learning continuous
representations of entities and relations tailored towards the link prediction
problem. Recent results indicate an ever increasing predictive ability of
current approaches on benchmark datasets. However, this effectiveness often
comes with the cost of over-parameterization and increased computationally
complexity. The former induces extensive hyperparameter optimization to
mitigate malicious overfitting. The latter magnifies the importance of winning
the hardware lottery. Here, we investigate a remedy for the first problem. We
propose a technique based on Kronecker decomposition to reduce the number of
parameters in a knowledge graph embedding model, while retaining its
expressiveness. Through Kronecker decomposition, large embedding matrices are
split into smaller embedding matrices during the training process. Hence,
embeddings of knowledge graphs are not plainly retrieved but reconstructed on
the fly. The decomposition ensures that elementwise interactions between three
embedding vectors are extended with interactions within each embedding vector.
This implicitly reduces redundancy in embedding vectors and encourages feature
reuse. To quantify the impact of applying Kronecker decomposition on embedding
matrices, we conduct a series of experiments on benchmark datasets. Our
experiments suggest that applying Kronecker decomposition on embedding matrices
leads to an improved parameter efficiency on all benchmark datasets. Moreover,
empirical evidence suggests that reconstructed embeddings entail robustness
against noise in the input knowledge graph. To foster reproducible research, we
provide an open-source implementation of our approach, including training and
evaluation scripts as well as pre-trained models in our knowledge graph
embedding framework (https://github.com/dice-group/dice-embeddings).
Related papers
- Efficient and Generalizable Certified Unlearning: A Hessian-free Recollection Approach [8.875278412741695]
Machine unlearning strives to uphold the data owners' right to be forgotten by enabling models to selectively forget specific data.
We develop an algorithm that achieves near-instantaneous unlearning as it only requires a vector addition operation.
arXiv Detail & Related papers (2024-04-02T07:54:18Z) - CONVERT:Contrastive Graph Clustering with Reliable Augmentation [110.46658439733106]
We propose a novel CONtrastiVe Graph ClustEring network with Reliable AugmenTation (CONVERT)
In our method, the data augmentations are processed by the proposed reversible perturb-recover network.
To further guarantee the reliability of semantics, a novel semantic loss is presented to constrain the network.
arXiv Detail & Related papers (2023-08-17T13:07:09Z) - Gradient-Based Spectral Embeddings of Random Dot Product Graphs [7.612218105739107]
In this paper, we show how to better solve the embedding problem of the Random Dot Product Graph (RDPG)
We develop a novel feasible optimization method in the resulting manifold.
Our open-source algorithm implementations are scalable, and unlike the they are robust to missing edge data and can track slowly, latent positions from streaming graphs.
arXiv Detail & Related papers (2023-07-25T21:09:55Z) - Boosting Differentiable Causal Discovery via Adaptive Sample Reweighting [62.23057729112182]
Differentiable score-based causal discovery methods learn a directed acyclic graph from observational data.
We propose a model-agnostic framework to boost causal discovery performance by dynamically learning the adaptive weights for the Reweighted Score function, ReScore.
arXiv Detail & Related papers (2023-03-06T14:49:59Z) - GraphLearner: Graph Node Clustering with Fully Learnable Augmentation [76.63963385662426]
Contrastive deep graph clustering (CDGC) leverages the power of contrastive learning to group nodes into different clusters.
We propose a Graph Node Clustering with Fully Learnable Augmentation, termed GraphLearner.
It introduces learnable augmentors to generate high-quality and task-specific augmented samples for CDGC.
arXiv Detail & Related papers (2022-12-07T10:19:39Z) - Fast and Provable Tensor Robust Principal Component Analysis via Scaled
Gradient Descent [30.299284742925852]
This paper tackles tensor robust principal component analysis (RPCA)
It aims to recover a low-rank tensor from its observations contaminated by sparse corruptions.
We show that the proposed algorithm achieves better and more scalable performance than state-of-the-art matrix and tensor RPCA algorithms.
arXiv Detail & Related papers (2022-06-18T04:01:32Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - Multilayer Clustered Graph Learning [66.94201299553336]
We use contrastive loss as a data fidelity term, in order to properly aggregate the observed layers into a representative graph.
Experiments show that our method leads to a clustered clusters w.r.t.
We learn a clustering algorithm for solving clustering problems.
arXiv Detail & Related papers (2020-10-29T09:58:02Z) - Embedding Expansion: Augmentation in Embedding Space for Deep Metric
Learning [17.19890778916312]
We propose an augmentation method in an embedding space for pair-based metric learning losses, called embedding expansion.
Because of its simplicity and flexibility, it can be used for existing metric learning losses without affecting model size, training speed, or optimization difficulty.
arXiv Detail & Related papers (2020-03-05T11:43:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.