Chasing Streams with Existential Rules
- URL: http://arxiv.org/abs/2205.02220v1
- Date: Wed, 4 May 2022 17:53:17 GMT
- Title: Chasing Streams with Existential Rules
- Authors: Jacopo Urbani, Markus Kr\"otzsch, Thomas Eiter
- Abstract summary: We study reasoning with existential rules to perform query answering over streams of data.
We extend LARS, a framework for rule-based stream reasoning, to support existential rules.
We show how to translate LARS with existentials into a semantics-preserving set of existential rules.
- Score: 18.660026838228625
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study reasoning with existential rules to perform query answering over
streams of data. On static databases, this problem has been widely studied, but
its extension to rapidly changing data has not yet been considered. To bridge
this gap, we extend LARS, a well-known framework for rule-based stream
reasoning, to support existential rules. For that, we show how to translate
LARS with existentials into a semantics-preserving set of existential rules. As
query answering with such rules is undecidable in general, we describe how to
leverage the temporal nature of streams and present suitable notions of
acyclicity that ensure decidability.
Related papers
- From SHAP to Rules: Distilling Expert Knowledge from Post-hoc Model Explanations in Time Series Classification [7.7491252992917445]
We propose a framework that converts numeric feature attributions from post-hoc, instance-wise explainers into structured, human-readable rules.<n>Our approach performs comparably to native rule-based methods like Anchor while scaling better to long TS and covering more instances.<n> Experiments on UCI datasets confirm that the resulting rule-based representations improve interpretability, decision transparency, and practical applicability for TS classification.
arXiv Detail & Related papers (2025-08-03T09:45:40Z) - Improve Rule Retrieval and Reasoning with Self-Induction and Relevance ReEstimate [11.890202757543307]
This paper systematically addresses the challenges of rule retrieval, a crucial yet underexplored area.<n> Vanilla retrieval methods using sparse or dense retrievers to directly search for relevant rules often suffer from low accuracy.<n>We propose Self-Induction Augmented Retrieval (SIAR), a novel approach that utilizes Large Language Models (LLMs) to induce potential inferential rules.
arXiv Detail & Related papers (2025-05-16T05:22:42Z) - Symbolic Working Memory Enhances Language Models for Complex Rule Application [87.34281749422756]
Large Language Models (LLMs) have shown remarkable reasoning performance but struggle with multi-step deductive reasoning.
We propose augmenting LLMs with external working memory and introduce a neurosymbolic framework for rule application.
Our framework iteratively performs symbolic rule grounding and LLM-based rule implementation.
arXiv Detail & Related papers (2024-08-24T19:11:54Z) - Conformance Checking of Fuzzy Logs against Declarative Temporal Specifications [22.272099472568893]
We consider the problem of checking whether fuzzy event data conform with declarative temporal rules specified as Declare patterns.
This requires relax the assumption that at each instant only one activity is executed.
We provide a proof-of-concept, efficient implementation based on the PyTorch Python library.
arXiv Detail & Related papers (2024-06-17T20:38:57Z) - Faithful Differentiable Reasoning with Reshuffled Region-based Embeddings [62.93577376960498]
Knowledge graph embedding methods learn geometric representations of entities and relations to predict plausible missing knowledge.<n>We propose RESHUFFLE, a model based on ordering constraints that can faithfully capture a much larger class of rule bases.<n>The entity embeddings in our framework can be learned by a Graph Neural Network (GNN), which effectively acts as a differentiable rule base.
arXiv Detail & Related papers (2024-06-13T18:37:24Z) - Can LLMs Reason with Rules? Logic Scaffolding for Stress-Testing and Improving LLMs [87.34281749422756]
Large language models (LLMs) have achieved impressive human-like performance across various reasoning tasks.
However, their mastery of underlying inferential rules still falls short of human capabilities.
We propose a logic scaffolding inferential rule generation framework, to construct an inferential rule base, ULogic.
arXiv Detail & Related papers (2024-02-18T03:38:51Z) - Chain-of-Table: Evolving Tables in the Reasoning Chain for Table
Understanding [79.9461269253121]
We propose the Chain-of-Table framework, where tabular data is explicitly used in the reasoning chain as a proxy for intermediate thoughts.
Chain-of-Table achieves new state-of-the-art performance on WikiTQ, FeTaQA, and TabFact benchmarks.
arXiv Detail & Related papers (2024-01-09T07:46:26Z) - ChatRule: Mining Logical Rules with Large Language Models for Knowledge
Graph Reasoning [107.61997887260056]
We propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs.
Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs.
To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs.
arXiv Detail & Related papers (2023-09-04T11:38:02Z) - Mitigating Catastrophic Forgetting in Task-Incremental Continual
Learning with Adaptive Classification Criterion [50.03041373044267]
We propose a Supervised Contrastive learning framework with adaptive classification criterion for Continual Learning.
Experiments show that CFL achieves state-of-the-art performance and has a stronger ability to overcome compared with the classification baselines.
arXiv Detail & Related papers (2023-05-20T19:22:40Z) - On the Foundations of Cycles in Bayesian Networks [4.312746668772342]
We present a foundational study regarding semantics for cyclic BNs that are generic and conservatively extend the cycle-free setting.
First, we propose constraint-based semantics that specify requirements for full joint distributions over a BN to be consistent with the local conditional probabilities and independencies.
Second, two kinds of limit semantics that formalize infinite unfolding approaches are introduced and shown to be computable by a Markov chain construction.
arXiv Detail & Related papers (2023-01-20T14:40:17Z) - RulE: Knowledge Graph Reasoning with Rule Embedding [69.31451649090661]
We propose a principled framework called textbfRulE (stands for Rule Embedding) to leverage logical rules to enhance KG reasoning.
RulE learns rule embeddings from existing triplets and first-order rules by jointly representing textbfentities, textbfrelations and textbflogical rules in a unified embedding space.
Results on multiple benchmarks reveal that our model outperforms the majority of existing embedding-based and rule-based approaches.
arXiv Detail & Related papers (2022-10-24T06:47:13Z) - Answering Fuzzy Queries over Fuzzy DL-Lite Ontologies [0.2741266294612776]
We study the problem of answering conjunctive queries and threshold queries in fuzzy DL-Lite.
For the idemdent G"odel t-norm, we provide an effective method based on a reduction to the classical case.
arXiv Detail & Related papers (2021-11-23T10:45:54Z) - Parallelisable Existential Rules: a Story of Pieces [2.20439695290991]
We introduce parallelisable sets of existential rules, for which the chase can be computed in a single breadth-first step from any instance.
We show that parallelisable rule sets are exactly those rule sets both bounded for the chase and belonging to a novel class of rules, called pieceful.
arXiv Detail & Related papers (2021-07-13T13:09:14Z) - Learning Implicitly with Noisy Data in Linear Arithmetic [94.66549436482306]
We extend implicit learning in PAC-Semantics to handle intervals and threshold uncertainty in the language of linear arithmetic.
We show that our implicit approach to learning optimal linear programming objective constraints significantly outperforms an explicit approach in practice.
arXiv Detail & Related papers (2020-10-23T19:08:46Z) - Fixed Point Semantics for Stream Reasoning [0.0]
Stream reasoning has emerged as a research area within the AI-community with many potential applications.
The rule-based formalism em LARS for non-monotonic stream reasoning under the answer set semantics has been introduced.
We show that our semantics is sound and constructive in the sense that answer sets are derivable bottom-up and free of circular justifications.
arXiv Detail & Related papers (2020-05-17T22:25:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.