Knowledge Sheaves: A Sheaf-Theoretic Framework for Knowledge Graph
Embedding
- URL: http://arxiv.org/abs/2110.03789v2
- Date: Sat, 18 Mar 2023 15:36:35 GMT
- Title: Knowledge Sheaves: A Sheaf-Theoretic Framework for Knowledge Graph
Embedding
- Authors: Thomas Gebhart, Jakob Hansen, Paul Schrater
- Abstract summary: We show that knowledge graph embedding is naturally expressed in the topological and categorical language of textitcellular sheaves
A knowledge graph embedding can be described as an approximate global section of an appropriate textitknowledge sheaf over the graph.
The resulting embeddings can be easily adapted for reasoning over composite relations without special training.
- Score: 1.5469452301122175
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Knowledge graph embedding involves learning representations of entities --
the vertices of the graph -- and relations -- the edges of the graph -- such
that the resulting representations encode the known factual information
represented by the knowledge graph and can be used in the inference of new
relations. We show that knowledge graph embedding is naturally expressed in the
topological and categorical language of \textit{cellular sheaves}: a knowledge
graph embedding can be described as an approximate global section of an
appropriate \textit{knowledge sheaf} over the graph, with consistency
constraints induced by the knowledge graph's schema. This approach provides a
generalized framework for reasoning about knowledge graph embedding models and
allows for the expression of a wide range of prior constraints on embeddings.
Further, the resulting embeddings can be easily adapted for reasoning over
composite relations without special training. We implement these ideas to
highlight the benefits of the extensions inspired by this new perspective.
Related papers
- From Latent to Lucid: Transforming Knowledge Graph Embeddings into Interpretable Structures [2.6451388057494283]
This paper introduces a post-hoc explainable AI method tailored for Knowledge Graph Embedding models.
Our approach directly decodes the latent representations encoded by Knowledge Graph Embedding models.
By identifying distinct structures within the subgraph neighborhoods of similarly embedded entities, our method translates these insights into human-understandable symbolic rules and facts.
arXiv Detail & Related papers (2024-06-03T19:54:11Z) - Extending Transductive Knowledge Graph Embedding Models for Inductive
Logical Relational Inference [0.5439020425819]
This work bridges the gap between traditional transductive knowledge graph embedding approaches and more recent inductive relation prediction models.
We introduce a generalized form of harmonic extension which leverages representations learned through transductive embedding methods to infer representations of new entities introduced at inference time as in the inductive setting.
In experiments on a number of large-scale knowledge graph embedding benchmarks, we find that this approach for extending the functionality of transductive knowledge graph embedding models is competitive with--and in some scenarios outperforms--several state-of-the-art models derived explicitly for such inductive tasks.
arXiv Detail & Related papers (2023-09-07T15:24:18Z) - InGram: Inductive Knowledge Graph Embedding via Relation Graphs [16.005051393690792]
In this paper, we propose an INductive knowledge GRAph eMbedding method, InGram, that can generate embeddings of new relations as well as new entities at inference time.
Experimental results show that InGram outperforms 14 different state-of-the-art methods on varied inductive learning scenarios.
arXiv Detail & Related papers (2023-05-31T16:10:42Z) - State of the Art and Potentialities of Graph-level Learning [54.68482109186052]
Graph-level learning has been applied to many tasks including comparison, regression, classification, and more.
Traditional approaches to learning a set of graphs rely on hand-crafted features, such as substructures.
Deep learning has helped graph-level learning adapt to the growing scale of graphs by extracting features automatically and encoding graphs into low-dimensional representations.
arXiv Detail & Related papers (2023-01-14T09:15:49Z) - KGLM: Integrating Knowledge Graph Structure in Language Models for Link
Prediction [0.0]
We introduce a new entity/relation embedding layer that learns to differentiate distinctive entity and relation types.
We show that further pre-training the language models with this additional embedding layer using the triples extracted from the knowledge graph, followed by the standard fine-tuning phase sets a new state-of-the-art performance for the link prediction task on the benchmark datasets.
arXiv Detail & Related papers (2022-11-04T20:38:12Z) - Joint Language Semantic and Structure Embedding for Knowledge Graph
Completion [66.15933600765835]
We propose to jointly embed the semantics in the natural language description of the knowledge triplets with their structure information.
Our method embeds knowledge graphs for the completion task via fine-tuning pre-trained language models.
Our experiments on a variety of knowledge graph benchmarks have demonstrated the state-of-the-art performance of our method.
arXiv Detail & Related papers (2022-09-19T02:41:02Z) - Conditional Attention Networks for Distilling Knowledge Graphs in
Recommendation [74.14009444678031]
We propose Knowledge-aware Conditional Attention Networks (KCAN) to incorporate knowledge graph into a recommender system.
We use a knowledge-aware attention propagation manner to obtain the node representation first, which captures the global semantic similarity on the user-item network and the knowledge graph.
Then, by applying a conditional attention aggregation on the subgraph, we refine the knowledge graph to obtain target-specific node representations.
arXiv Detail & Related papers (2021-11-03T09:40:43Z) - Knowledge Graph Completion with Text-aided Regularization [2.8361571014635407]
Knowledge Graph Completion is a task of expanding the knowledge graph/base through estimating possible entities.
Traditional approaches mainly focus on using the existing graphical information that is intrinsic of the graph.
We try numerous ways of using extracted or raw textual information to help existing KG embedding frameworks reach better prediction results.
arXiv Detail & Related papers (2021-01-22T06:10:09Z) - Structural Landmarking and Interaction Modelling: on Resolution Dilemmas
in Graph Classification [50.83222170524406]
We study the intrinsic difficulty in graph classification under the unified concept of resolution dilemmas''
We propose SLIM'', an inductive neural network model for Structural Landmarking and Interaction Modelling.
arXiv Detail & Related papers (2020-06-29T01:01:42Z) - Generative Adversarial Zero-Shot Relational Learning for Knowledge
Graphs [96.73259297063619]
We consider a novel formulation, zero-shot learning, to free this cumbersome curation.
For newly-added relations, we attempt to learn their semantic features from their text descriptions.
We leverage Generative Adrial Networks (GANs) to establish the connection between text and knowledge graph domain.
arXiv Detail & Related papers (2020-01-08T01:19:08Z) - Bridging Knowledge Graphs to Generate Scene Graphs [49.69377653925448]
We propose a novel graph-based neural network that iteratively propagates information between the two graphs, as well as within each of them.
Our Graph Bridging Network, GB-Net, successively infers edges and nodes, allowing to simultaneously exploit and refine the rich, heterogeneous structure of the interconnected scene and commonsense graphs.
arXiv Detail & Related papers (2020-01-07T23:35:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.