Cleaning Inconsistent Data in Temporal DL-Lite Under Best Repair
Semantics
- URL: http://arxiv.org/abs/2108.12149v2
- Date: Mon, 30 Aug 2021 07:21:14 GMT
- Title: Cleaning Inconsistent Data in Temporal DL-Lite Under Best Repair
Semantics
- Authors: Mourad Ouziri (LIPADE - EA 2517), Sabiha Tahrat (LIPADE - EA 2517),
Salima Benbernou (LIPADE - EA 2517), Mourad Ouzirri
- Abstract summary: We address the problem of handling inconsistent data in Temporal Description Logic (TDL) knowledge bases.
Considering the data part of the Knowledge Base as the source of inconsistency over time, we propose an ABox repair approach.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the problem of handling inconsistent data in
Temporal Description Logic (TDL) knowledge bases. Considering the data part of
the Knowledge Base as the source of inconsistency over time, we propose an ABox
repair approach. This is the first work handling the repair in TDL Knowledge
bases. To do so, our goal is twofold: 1) detect temporal inconsistencies and 2)
propose a data temporal reparation. For the inconsistency detection, we propose
a reduction approach from TDL to DL which allows to provide a tight NP-complete
upper bound for TDL concept satisfiability and to use highly optimised DL
reasoners that can bring precise explanation (the set of inconsistent data
assertions). Thereafter, from the obtained explanation, we propose a method for
automatically computing the best repair in the temporal setting based on the
allowed rigid predicates and the time order of assertions.
Related papers
- Disperse-Then-Merge: Pushing the Limits of Instruction Tuning via Alignment Tax Reduction [75.25114727856861]
Large language models (LLMs) tend to suffer from deterioration at the latter stage ofSupervised fine-tuning process.
We introduce a simple disperse-then-merge framework to address the issue.
Our framework outperforms various sophisticated methods such as data curation and training regularization on a series of standard knowledge and reasoning benchmarks.
arXiv Detail & Related papers (2024-05-22T08:18:19Z) - Fact Checking Beyond Training Set [64.88575826304024]
We show that the retriever-reader suffers from performance deterioration when it is trained on labeled data from one domain and used in another domain.
We propose an adversarial algorithm to make the retriever component robust against distribution shift.
We then construct eight fact checking scenarios from these datasets, and compare our model to a set of strong baseline models.
arXiv Detail & Related papers (2024-03-27T15:15:14Z) - Dated Data: Tracing Knowledge Cutoffs in Large Language Models [47.987664966633865]
We propose a simple approach to estimate effective cutoffs on the resource-level temporal alignment of an LLM.
We find that effective cutoffs often differ from reported cutoffs.
Our analysis reveals two reasons for these inconsistencies: (1) temporal biases of CommonCrawl data due to non-trivial amounts of old data in new dumps and (2) complications in LLM deduplication schemes involving semantic duplicates and lexical near-duplicates.
arXiv Detail & Related papers (2024-03-19T17:57:58Z) - DAGnosis: Localized Identification of Data Inconsistencies using
Structures [73.39285449012255]
Identification and appropriate handling of inconsistencies in data at deployment time is crucial to reliably use machine learning models.
We use directed acyclic graphs (DAGs) to encode the training set's features probability distribution and independencies as a structure.
Our method, called DAGnosis, leverages these structural interactions to bring valuable and insightful data-centric conclusions.
arXiv Detail & Related papers (2024-02-26T11:29:16Z) - Better patching using LLM prompting, via Self-Consistency [5.892272127970584]
Self-consistency (S-C) is an exciting, substantially better technique for generating explanations for problems.
This paper describes an application of the S-C approach to program repair, using the commit log on the fix as the explanation.
We achieve state-of-the art results, beating previous approaches to prompting-based program repair on the MODIT dataset.
arXiv Detail & Related papers (2023-05-31T18:28:46Z) - Towards Bridging the Gaps between the Right to Explanation and the Right
to be Forgotten [14.636997283608414]
The right to explanation allows individuals to request an actionable explanation for an algorithmic decision.
The right to be forgotten grants them the right to ask for their data to be deleted from all the databases and models of an organization.
We propose the first algorithmic framework to resolve the tension between the two principles.
arXiv Detail & Related papers (2023-02-08T19:03:00Z) - Supporting Vision-Language Model Inference with Confounder-pruning Knowledge Prompt [71.77504700496004]
Vision-language models are pre-trained by aligning image-text pairs in a common space to deal with open-set visual concepts.
To boost the transferability of the pre-trained models, recent works adopt fixed or learnable prompts.
However, how and what prompts can improve inference performance remains unclear.
arXiv Detail & Related papers (2022-05-23T07:51:15Z) - Second-Order Specifications and Quantifier Elimination for Consistent
Query Answering in Databases [0.0]
We show how to use the repair programs to transform the problem of consistent query answering into a problem of reasoning w.r.t.
It also investigated how a first-order theory can be obtained instead by applying second-order quantifier elimination techniques.
arXiv Detail & Related papers (2021-08-19T01:06:01Z) - Automated Reasoning in Temporal DL-Lite [65.9825143048822]
This paper investigates the feasibility of automated reasoning over temporal DL-Lite (TDL-Lite) knowledge bases (KBs)
We test the usage of off-theshelf reasoners to check satisfiability of TDL-Lite KBs.
In an effort to make the usage of TDL-Lite KBs a reality, we present a fully fledged tool with a graphical interface to design them.
arXiv Detail & Related papers (2020-08-17T16:40:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.