Building Rule Hierarchies for Efficient Logical Rule Learning from
Knowledge Graphs
- URL: http://arxiv.org/abs/2006.16171v3
- Date: Thu, 22 Oct 2020 23:32:53 GMT
- Title: Building Rule Hierarchies for Efficient Logical Rule Learning from
Knowledge Graphs
- Authors: Yulong Gu, Yu Guan, Paolo Missier
- Abstract summary: We propose new methods for pruning unpromising rules using rule hierarchies.
We show that the application of HPMs is effective in removing unpromising rules.
- Score: 20.251630903853016
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many systems have been developed in recent years to mine logical rules from
large-scale Knowledge Graphs (KGs), on the grounds that representing
regularities as rules enables both the interpretable inference of new facts,
and the explanation of known facts. Among these systems, the walk-based methods
that generate the instantiated rules containing constants by abstracting
sampled paths in KGs demonstrate strong predictive performance and
expressivity. However, due to the large volume of possible rules, these systems
do not scale well where computational resources are often wasted on generating
and evaluating unpromising rules. In this work, we address such scalability
issues by proposing new methods for pruning unpromising rules using rule
hierarchies. The approach consists of two phases. Firstly, since rule
hierarchies are not readily available in walk-based methods, we have built a
Rule Hierarchy Framework (RHF), which leverages a collection of subsumption
frameworks to build a proper rule hierarchy from a set of learned rules. And
secondly, we adapt RHF to an existing rule learner where we design and
implement two methods for Hierarchical Pruning (HPMs), which utilize the
generated hierarchies to remove irrelevant and redundant rules. Through
experiments over four public benchmark datasets, we show that the application
of HPMs is effective in removing unpromising rules, which leads to significant
reductions in the runtime as well as in the number of learned rules, without
compromising the predictive performance.
Related papers
- A Scalable Matrix Visualization for Understanding Tree Ensemble Classifiers [20.416696003269674]
This paper introduces a scalable visual analysis method to explain tree ensemble classifiers that contain tens of thousands of rules.
We develop an anomaly-biased model reduction method to prioritize these rules at each hierarchical level.
Our method fosters a deeper understanding of both common and anomalous rules, thereby enhancing interpretability without sacrificing comprehensiveness.
arXiv Detail & Related papers (2024-09-05T01:48:11Z) - Symbolic Working Memory Enhances Language Models for Complex Rule Application [87.34281749422756]
Large Language Models (LLMs) have shown remarkable reasoning performance but struggle with multi-step deductive reasoning.
We propose augmenting LLMs with external working memory and introduce a neurosymbolic framework for rule application.
Our framework iteratively performs symbolic rule grounding and LLM-based rule implementation.
arXiv Detail & Related papers (2024-08-24T19:11:54Z) - Can LLMs Reason with Rules? Logic Scaffolding for Stress-Testing and Improving LLMs [87.34281749422756]
Large language models (LLMs) have achieved impressive human-like performance across various reasoning tasks.
However, their mastery of underlying inferential rules still falls short of human capabilities.
We propose a logic scaffolding inferential rule generation framework, to construct an inferential rule base, ULogic.
arXiv Detail & Related papers (2024-02-18T03:38:51Z) - ChatRule: Mining Logical Rules with Large Language Models for Knowledge
Graph Reasoning [107.61997887260056]
We propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs.
Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs.
To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs.
arXiv Detail & Related papers (2023-09-04T11:38:02Z) - Knowledge Reasoning via Jointly Modeling Knowledge Graphs and Soft Rules [17.301284626706856]
Methods of knowledge graph completion (KGC) can be classified into two major categories: rule-based reasoning and embedding-based reasoning.
We propose a novel method that injects rules and learns representations iteratively to take full advantage of rules and embeddings.
arXiv Detail & Related papers (2023-01-07T05:24:29Z) - RulE: Knowledge Graph Reasoning with Rule Embedding [69.31451649090661]
We propose a principled framework called textbfRulE (stands for Rule Embedding) to leverage logical rules to enhance KG reasoning.
RulE learns rule embeddings from existing triplets and first-order rules by jointly representing textbfentities, textbfrelations and textbflogical rules in a unified embedding space.
Results on multiple benchmarks reveal that our model outperforms the majority of existing embedding-based and rule-based approaches.
arXiv Detail & Related papers (2022-10-24T06:47:13Z) - Distilling Task-specific Logical Rules from Large Pre-trained Models [24.66436804853525]
We develop a novel framework to distill task-specific logical rules from large pre-trained models.
Specifically, we borrow recent prompt-based language models as the knowledge expert to yield initial seed rules.
Experiments on three public named entity tagging benchmarks demonstrate the effectiveness of our proposed framework.
arXiv Detail & Related papers (2022-10-06T09:12:18Z) - Rewriting a Deep Generative Model [56.91974064348137]
We introduce a new problem setting: manipulation of specific rules encoded by a deep generative model.
We propose a formulation in which the desired rule is changed by manipulating a layer of a deep network as a linear associative memory.
We present a user interface to enable users to interactively change the rules of a generative model to achieve desired effects.
arXiv Detail & Related papers (2020-07-30T17:58:16Z) - Towards Learning Instantiated Logical Rules from Knowledge Graphs [20.251630903853016]
We present GPFL, a probabilistic learner rule optimized to mine instantiated first-order logic rules from knowledge graphs.
GPFL utilizes a novel two-stage rule generation mechanism that first generalizes extracted paths into templates that are acyclic abstract rules.
We reveal the presence of overfitting rules, their impact on the predictive performance, and the effectiveness of a simple validation method filtering out overfitting rules.
arXiv Detail & Related papers (2020-03-13T00:32:46Z) - Hierarchical Variational Imitation Learning of Control Programs [131.7671843857375]
We propose a variational inference method for imitation learning of a control policy represented by parametrized hierarchical procedures (PHP)
Our method discovers the hierarchical structure in a dataset of observation-action traces of teacher demonstrations, by learning an approximate posterior distribution over the latent sequence of procedure calls and terminations.
We demonstrate a novel benefit of variational inference in the context of hierarchical imitation learning: in decomposing the policy into simpler procedures, inference can leverage acausal information that is unused by other methods.
arXiv Detail & Related papers (2019-12-29T08:57:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.