RulE: Knowledge Graph Reasoning with Rule Embedding
- URL: http://arxiv.org/abs/2210.14905v3
- Date: Mon, 20 May 2024 11:21:40 GMT
- Title: RulE: Knowledge Graph Reasoning with Rule Embedding
- Authors: Xiaojuan Tang, Song-Chun Zhu, Yitao Liang, Muhan Zhang,
- Abstract summary: We propose a principled framework called textbfRulE (stands for Rule Embedding) to leverage logical rules to enhance KG reasoning.
RulE learns rule embeddings from existing triplets and first-order rules by jointly representing textbfentities, textbfrelations and textbflogical rules in a unified embedding space.
Results on multiple benchmarks reveal that our model outperforms the majority of existing embedding-based and rule-based approaches.
- Score: 69.31451649090661
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Knowledge graph (KG) reasoning is an important problem for knowledge graphs. In this paper, we propose a novel and principled framework called \textbf{RulE} (stands for {Rul}e {E}mbedding) to effectively leverage logical rules to enhance KG reasoning. Unlike knowledge graph embedding (KGE) methods, RulE learns rule embeddings from existing triplets and first-order {rules} by jointly representing \textbf{entities}, \textbf{relations} and \textbf{logical rules} in a unified embedding space. Based on the learned rule embeddings, a confidence score can be calculated for each rule, reflecting its consistency with the observed triplets. This allows us to perform logical rule inference in a soft way, thus alleviating the brittleness of logic. On the other hand, RulE injects prior logical rule information into the embedding space, enriching and regularizing the entity/relation embeddings. This makes KGE alone perform better too. RulE is conceptually simple and empirically effective. We conduct extensive experiments to verify each component of RulE. Results on multiple benchmarks reveal that our model outperforms the majority of existing embedding-based and rule-based approaches.
Related papers
- Can LLMs Reason with Rules? Logic Scaffolding for Stress-Testing and Improving LLMs [87.34281749422756]
Large language models (LLMs) have achieved impressive human-like performance across various reasoning tasks.
However, their mastery of underlying inferential rules still falls short of human capabilities.
We propose a logic scaffolding inferential rule generation framework, to construct an inferential rule base, ULogic.
arXiv Detail & Related papers (2024-02-18T03:38:51Z) - ChatRule: Mining Logical Rules with Large Language Models for Knowledge
Graph Reasoning [107.61997887260056]
We propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs.
Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs.
To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs.
arXiv Detail & Related papers (2023-09-04T11:38:02Z) - Simple Rule Injection for ComplEx Embeddings [7.19573352891936]
In this work, we propose InjEx, a mechanism to inject multiple types of rules through simple constraints.
We demonstrate that InjEx infuses interpretable prior knowledge into the embedding space.
Our experimental results reveal that InjEx outperforms both baseline KGC models as well as specialized few-shot models.
arXiv Detail & Related papers (2023-08-07T03:19:59Z) - Logical Entity Representation in Knowledge-Graphs for Differentiable
Rule Learning [71.05093203007357]
We propose Logical Entity RePresentation (LERP) to encode contextual information of entities in the knowledge graph.
A LERP is designed as a vector of probabilistic logical functions on the entity's neighboring sub-graph.
Our model outperforms other rule learning methods in knowledge graph completion and is comparable or even superior to state-of-the-art black-box methods.
arXiv Detail & Related papers (2023-05-22T05:59:22Z) - Knowledge Reasoning via Jointly Modeling Knowledge Graphs and Soft Rules [17.301284626706856]
Methods of knowledge graph completion (KGC) can be classified into two major categories: rule-based reasoning and embedding-based reasoning.
We propose a novel method that injects rules and learns representations iteratively to take full advantage of rules and embeddings.
arXiv Detail & Related papers (2023-01-07T05:24:29Z) - Combining Rules and Embeddings via Neuro-Symbolic AI for Knowledge Base
Completion [59.093293389123424]
We show that not all rule-based Knowledge Base Completion models are the same.
We propose two distinct approaches that learn in one case: 1) a mixture of relations and the other 2) a mixture of paths.
When implemented on top of neuro-symbolic AI, which learns rules by extending Boolean logic to real-valued logic, the latter model leads to superior KBC accuracy outperforming state-of-the-art rule-based KBC by 2-10% in terms of mean reciprocal rank.
arXiv Detail & Related papers (2021-09-16T17:54:56Z) - Theoretical Rule-based Knowledge Graph Reasoning by Connectivity
Dependency Discovery [2.945948598480997]
We present a theory for rule-based knowledge graph reasoning, based on which the connectivity dependencies in the graph are captured via multiple rule types.
Results show that our RuleDict model not only provides precise rules to interpret new triples, but also achieves state-of-the-art performances on one benchmark knowledge graph completion task.
arXiv Detail & Related papers (2020-11-12T03:00:20Z) - RNNLogic: Learning Logic Rules for Reasoning on Knowledge Graphs [91.71504177786792]
This paper studies learning logic rules for reasoning on knowledge graphs.
Logic rules provide interpretable explanations when used for prediction as well as being able to generalize to other tasks.
Existing methods either suffer from the problem of searching in a large search space or ineffective optimization due to sparse rewards.
arXiv Detail & Related papers (2020-10-08T14:47:02Z) - EM-RBR: a reinforced framework for knowledge graph completion from
reasoning perspective [3.6188659868203397]
We propose a general framework, named EM-RBR, capable of combining the advantages of reasoning based on rules and the state-of-the-art models of embedding.
EM-RBR aims to utilize relational background knowledge contained in rules to conduct multi-relation reasoning link prediction.
In experiments, we demonstrate that EM-RBR achieves better performance compared with previous models on FB15k, WN18 and our new dataset FB15k-R.
arXiv Detail & Related papers (2020-09-18T07:02:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.