An Experiment in Retrofitting Competency Questions for Existing
Ontologies
- URL: http://arxiv.org/abs/2311.05662v1
- Date: Thu, 9 Nov 2023 08:57:39 GMT
- Title: An Experiment in Retrofitting Competency Questions for Existing
Ontologies
- Authors: Reham Alharbi and Valentina Tamma and Floriana Grasso and Terry Payne
- Abstract summary: Inspecting CQs together with the axioms provides critical insights into the scope and applicability of the CQs.
CQs are integral to the majority of engineering methodologies, but the practice of publishing CQs alongside the on artefacts is not widely observed.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Competency Questions (CQs) are a form of ontology functional requirements
expressed as natural language questions. Inspecting CQs together with the
axioms in an ontology provides critical insights into the intended scope and
applicability of the ontology. CQs also underpin a number of tasks in the
development of ontologies e.g. ontology reuse, ontology testing, requirement
specification, and the definition of patterns that implement such requirements.
Although CQs are integral to the majority of ontology engineering
methodologies, the practice of publishing CQs alongside the ontological
artefacts is not widely observed by the community. In this context, we present
an experiment in retrofitting CQs from existing ontologies. We propose
RETROFIT-CQs, a method to extract candidate CQs directly from ontologies using
Generative AI. In the paper we present the pipeline that facilitates the
extraction of CQs by leveraging Large Language Models (LLMs) and we discuss its
application to a number of existing ontologies.
Related papers
- A RAG Approach for Generating Competency Questions in Ontology Engineering [1.0044270899550196]
With the emergence of Large Language Models (LLMs), there arises the possibility to automate and enhance this process.
We present a retrieval-augmented generation (RAG) approach that uses LLMs for the automatic generation of CQs.
We conduct experiments using GPT-4 on two domain engineering tasks and compare results against ground-truth CQs constructed by domain experts.
arXiv Detail & Related papers (2024-09-13T13:34:32Z) - Controlled Query Evaluation through Epistemic Dependencies [7.502796412126707]
We show the expressive abilities of our framework and study the data complexity of CQE for (unions of) conjunctive queries.
We prove tractability for the case of acyclic dependencies by providing a suitable query algorithm.
arXiv Detail & Related papers (2024-05-03T19:48:07Z) - Structural generalization is hard for sequence-to-sequence models [85.0087839979613]
Sequence-to-sequence (seq2seq) models have been successful across many NLP tasks.
Recent work on compositional generalization has shown that seq2seq models achieve very low accuracy in generalizing to linguistic structures that were not seen in training.
arXiv Detail & Related papers (2022-10-24T09:03:03Z) - ProQA: Structural Prompt-based Pre-training for Unified Question
Answering [84.59636806421204]
ProQA is a unified QA paradigm that solves various tasks through a single model.
It concurrently models the knowledge generalization for all QA tasks while keeping the knowledge customization for every specific QA task.
ProQA consistently boosts performance on both full data fine-tuning, few-shot learning, and zero-shot testing scenarios.
arXiv Detail & Related papers (2022-05-09T04:59:26Z) - Finite Entailment of UCRPQs over ALC Ontologies [0.82179684645881]
We consider the query language, unions of conjunctive regular path queries (UCRPQs)
We show a tight 2EXP bound for entailment of UCRPQs using the description logic ALC.
There is a novel automata-based technique introducing a stratification of interpretations induced by the deterministic finite automaton underlying the input automaton.
arXiv Detail & Related papers (2022-04-29T17:38:13Z) - Improving Unsupervised Question Answering via Summarization-Informed
Question Generation [47.96911338198302]
Question Generation (QG) is the task of generating a plausible question for a passage, answer> pair.
We make use of freely available news summary data, transforming declarative sentences into appropriate questions using dependency parsing, named entity recognition and semantic role labeling.
The resulting questions are then combined with the original news articles to train an end-to-end neural QG model.
arXiv Detail & Related papers (2021-09-16T13:08:43Z) - BigCQ: A large-scale synthetic dataset of competency question patterns
formalized into SPARQL-OWL query templates [0.06445605125467574]
BigCQ is the biggest dataset of CQ templates with their formalizations into SPARQL-OWL query templates.
We describe the dataset in detail, provide a description of the process leading to the creation of the dataset and analyze how well the dataset covers real-world examples.
arXiv Detail & Related papers (2021-05-20T07:59:59Z) - Retrieving and Reading: A Comprehensive Survey on Open-domain Question
Answering [62.88322725956294]
We review the latest research trends in OpenQA, with particular attention to systems that incorporate neural MRC techniques.
We introduce modern OpenQA architecture named Retriever-Reader'' and analyze the various systems that follow this architecture.
We then discuss key challenges to developing OpenQA systems and offer an analysis of benchmarks that are commonly used.
arXiv Detail & Related papers (2021-01-04T04:47:46Z) - Question Answering over Knowledge Bases by Leveraging Semantic Parsing
and Neuro-Symbolic Reasoning [73.00049753292316]
We propose a semantic parsing and reasoning-based Neuro-Symbolic Question Answering(NSQA) system.
NSQA achieves state-of-the-art performance on QALD-9 and LC-QuAD 1.0.
arXiv Detail & Related papers (2020-12-03T05:17:55Z) - A Review of Uncertainty Quantification in Deep Learning: Techniques,
Applications and Challenges [76.20963684020145]
Uncertainty quantification (UQ) plays a pivotal role in reduction of uncertainties during both optimization and decision making processes.
Bizarre approximation and ensemble learning techniques are two most widely-used UQ methods in the literature.
This study reviews recent advances in UQ methods used in deep learning and investigates the application of these methods in reinforcement learning.
arXiv Detail & Related papers (2020-11-12T06:41:05Z) - CliniQG4QA: Generating Diverse Questions for Domain Adaptation of
Clinical Question Answering [27.45623324582005]
Clinical question answering (QA) aims to automatically answer questions from medical professionals based on clinical texts.
We propose CliniQG4QA, which leverages question generation (QG) to synthesize QA pairs on new clinical contexts.
In order to generate diverse types of questions that are essential for training QA models, we introduce a seq2seq-based question phrase prediction (QPP) module.
arXiv Detail & Related papers (2020-10-30T02:06:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.