Extending Transductive Knowledge Graph Embedding Models for Inductive
Logical Relational Inference
- URL: http://arxiv.org/abs/2309.03773v1
- Date: Thu, 7 Sep 2023 15:24:18 GMT
- Title: Extending Transductive Knowledge Graph Embedding Models for Inductive
Logical Relational Inference
- Authors: Thomas Gebhart and John Cobb
- Abstract summary: This work bridges the gap between traditional transductive knowledge graph embedding approaches and more recent inductive relation prediction models.
We introduce a generalized form of harmonic extension which leverages representations learned through transductive embedding methods to infer representations of new entities introduced at inference time as in the inductive setting.
In experiments on a number of large-scale knowledge graph embedding benchmarks, we find that this approach for extending the functionality of transductive knowledge graph embedding models is competitive with--and in some scenarios outperforms--several state-of-the-art models derived explicitly for such inductive tasks.
- Score: 0.5439020425819
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many downstream inference tasks for knowledge graphs, such as relation
prediction, have been handled successfully by knowledge graph embedding
techniques in the transductive setting. To address the inductive setting
wherein new entities are introduced into the knowledge graph at inference time,
more recent work opts for models which learn implicit representations of the
knowledge graph through a complex function of a network's subgraph structure,
often parametrized by graph neural network architectures. These come at the
cost of increased parametrization, reduced interpretability and limited
generalization to other downstream inference tasks. In this work, we bridge the
gap between traditional transductive knowledge graph embedding approaches and
more recent inductive relation prediction models by introducing a generalized
form of harmonic extension which leverages representations learned through
transductive embedding methods to infer representations of new entities
introduced at inference time as in the inductive setting. This harmonic
extension technique provides the best such approximation, can be implemented
via an efficient iterative scheme, and can be employed to answer a family of
conjunctive logical queries over the knowledge graph, further expanding the
capabilities of transductive embedding methods. In experiments on a number of
large-scale knowledge graph embedding benchmarks, we find that this approach
for extending the functionality of transductive knowledge graph embedding
models to perform knowledge graph completion and answer logical queries in the
inductive setting is competitive with--and in some scenarios
outperforms--several state-of-the-art models derived explicitly for such
inductive tasks.
Related papers
- Introducing Diminutive Causal Structure into Graph Representation Learning [19.132025125620274]
We introduce a novel method that enables Graph Neural Networks (GNNs) to glean insights from specialized diminutive causal structures.
Our method specifically extracts causal knowledge from the model representation of these diminutive causal structures.
arXiv Detail & Related papers (2024-06-13T00:18:20Z) - From Latent to Lucid: Transforming Knowledge Graph Embeddings into Interpretable Structures [2.6451388057494283]
This paper introduces a post-hoc explainable AI method tailored for Knowledge Graph Embedding models.
Our approach directly decodes the latent representations encoded by Knowledge Graph Embedding models.
By identifying distinct structures within the subgraph neighborhoods of similarly embedded entities, our method translates these insights into human-understandable symbolic rules and facts.
arXiv Detail & Related papers (2024-06-03T19:54:11Z) - Rethinking Dimensional Rationale in Graph Contrastive Learning from Causal Perspective [15.162584339143239]
Graph contrastive learning is a general learning paradigm excelling at capturing invariant information from diverse perturbations in graphs.
Recent works focus on exploring the structural rationale from graphs, thereby increasing the discriminability of the invariant information.
We propose the dimensional rationale-aware graph contrastive learning approach, which introduces a learnable dimensional rationale acquiring network and a redundancy reduction constraint.
arXiv Detail & Related papers (2023-12-16T10:05:18Z) - Graph Condensation for Inductive Node Representation Learning [59.76374128436873]
We propose mapping-aware graph condensation (MCond)
MCond integrates new nodes into the synthetic graph for inductive representation learning.
On the Reddit dataset, MCond achieves up to 121.5x inference speedup and 55.9x reduction in storage requirements.
arXiv Detail & Related papers (2023-07-29T12:11:14Z) - Towards Few-shot Inductive Link Prediction on Knowledge Graphs: A
Relational Anonymous Walk-guided Neural Process Approach [49.00753238429618]
Few-shot inductive link prediction on knowledge graphs aims to predict missing links for unseen entities with few-shot links observed.
Recent inductive methods utilize the sub-graphs around unseen entities to obtain the semantics and predict links inductively.
We propose a novel relational anonymous walk-guided neural process for few-shot inductive link prediction on knowledge graphs, denoted as RawNP.
arXiv Detail & Related papers (2023-06-26T12:02:32Z) - Message Intercommunication for Inductive Relation Reasoning [49.731293143079455]
We develop a novel inductive relation reasoning model called MINES.
We introduce a Message Intercommunication mechanism on the Neighbor-Enhanced Subgraph.
Our experiments show that MINES outperforms existing state-of-the-art models.
arXiv Detail & Related papers (2023-05-23T13:51:46Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Knowledge Sheaves: A Sheaf-Theoretic Framework for Knowledge Graph
Embedding [1.5469452301122175]
We show that knowledge graph embedding is naturally expressed in the topological and categorical language of textitcellular sheaves
A knowledge graph embedding can be described as an approximate global section of an appropriate textitknowledge sheaf over the graph.
The resulting embeddings can be easily adapted for reasoning over composite relations without special training.
arXiv Detail & Related papers (2021-10-07T20:54:40Z) - Quantifying Challenges in the Application of Graph Representation
Learning [0.0]
We provide an application oriented perspective to a set of popular embedding approaches.
We evaluate their representational power with respect to real-world graph properties.
Our results suggest that "one-to-fit-all" GRL approaches are hard to define in real-world scenarios.
arXiv Detail & Related papers (2020-06-18T03:19:43Z) - Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph
Link Prediction [69.1473775184952]
We introduce a realistic problem of few-shot out-of-graph link prediction.
We tackle this problem with a novel transductive meta-learning framework.
We validate our model on multiple benchmark datasets for knowledge graph completion and drug-drug interaction prediction.
arXiv Detail & Related papers (2020-06-11T17:42:46Z) - Graph Representation Learning via Graphical Mutual Information
Maximization [86.32278001019854]
We propose a novel concept, Graphical Mutual Information (GMI), to measure the correlation between input graphs and high-level hidden representations.
We develop an unsupervised learning model trained by maximizing GMI between the input and output of a graph neural encoder.
arXiv Detail & Related papers (2020-02-04T08:33:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.