Efficient Contextualization using Top-k Operators for Question Answering
over Knowledge Graphs
- URL: http://arxiv.org/abs/2108.08597v2
- Date: Sat, 21 Aug 2021 09:23:56 GMT
- Title: Efficient Contextualization using Top-k Operators for Question Answering
over Knowledge Graphs
- Authors: Philipp Christmann, Rishiraj Saha Roy, Gerhard Weikum
- Abstract summary: This work presents ECQA, an efficient method that prunes irrelevant parts of the search space using KB-aware signals.
Experiments with two recent QA benchmarks demonstrate the superiority of ECQA over state-of-the-art baselines with respect to answer presence, size of the search space, and runtimes.
- Score: 24.520002698010856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Answering complex questions over knowledge bases (KB-QA) faces huge input
data with billions of facts, involving millions of entities and thousands of
predicates. For efficiency, QA systems first reduce the answer search space by
identifying a set of facts that is likely to contain all answers and relevant
cues. The most common technique is to apply named entity disambiguation (NED)
systems to the question, and retrieve KB facts for the disambiguated entities.
This work presents ECQA, an efficient method that prunes irrelevant parts of
the search space using KB-aware signals. ECQA is based on top-k query
processing over score-ordered lists of KB items that combine signals about
lexical matching, relevance to the question, coherence among candidate items,
and connectivity in the KB graph. Experiments with two recent QA benchmarks
demonstrate the superiority of ECQA over state-of-the-art baselines with
respect to answer presence, size of the search space, and runtimes.
Related papers
- MarkQA: A large scale KBQA dataset with numerical reasoning [11.072552105311484]
We propose a new task, NR-KBQA, which requires the ability to perform both multi-hop reasoning and numerical reasoning.
We design a logic form in Python format called PyQL to represent the reasoning process of numerical reasoning questions.
We present a large dataset called MarkQA, which is automatically constructed from a small set of seeds.
arXiv Detail & Related papers (2023-10-24T04:50:59Z) - Two is Better Than One: Answering Complex Questions by Multiple
Knowledge Sources with Generalized Links [31.941956320431217]
We formulate the novel Multi-KB-QA task that leverages the full and partial links among multiple KBs to derive correct answers.
We propose a method for Multi-KB-QA that encodes all link relations in the KB embedding to score and rank candidate answers.
arXiv Detail & Related papers (2023-09-11T02:31:41Z) - BigText-QA: Question Answering over a Large-Scale Hybrid Knowledge Graph [23.739432128095107]
BigText-QA is able to answer questions based on a structured knowledge graph.
Our results demonstrate that BigText-QA outperforms DrQA, a neural-network-based QA system, and achieves competitive results to QUEST, a graph-based unsupervised QA system.
arXiv Detail & Related papers (2022-12-12T09:49:02Z) - UniKGQA: Unified Retrieval and Reasoning for Solving Multi-hop Question
Answering Over Knowledge Graph [89.98762327725112]
Multi-hop Question Answering over Knowledge Graph(KGQA) aims to find the answer entities that are multiple hops away from the topic entities mentioned in a natural language question.
We propose UniKGQA, a novel approach for multi-hop KGQA task, by unifying retrieval and reasoning in both model architecture and parameter learning.
arXiv Detail & Related papers (2022-12-02T04:08:09Z) - TIARA: Multi-grained Retrieval for Robust Question Answering over Large
Knowledge Bases [20.751369684593985]
TIARA outperforms previous SOTA, including those using PLMs or oracle entity annotations, by at least 4.1 and 1.1 F1 points on GrailQA and WebQuestionsSP.
arXiv Detail & Related papers (2022-10-24T02:41:10Z) - Knowledge Base Question Answering by Case-based Reasoning over Subgraphs [81.22050011503933]
We show that our model answers queries requiring complex reasoning patterns more effectively than existing KG completion algorithms.
The proposed model outperforms or performs competitively with state-of-the-art models on several KBQA benchmarks.
arXiv Detail & Related papers (2022-02-22T01:34:35Z) - A Benchmark for Generalizable and Interpretable Temporal Question
Answering over Knowledge Bases [67.33560134350427]
TempQA-WD is a benchmark dataset for temporal reasoning.
It is based on Wikidata, which is the most frequently curated, openly available knowledge base.
arXiv Detail & Related papers (2022-01-15T08:49:09Z) - Effective FAQ Retrieval and Question Matching With Unsupervised
Knowledge Injection [10.82418428209551]
We propose a contextual language model for retrieving appropriate answers to frequently asked questions.
We also explore to capitalize on domain-specific topically-relevant relations between words in an unsupervised manner.
We evaluate variants of our approach on a publicly-available Chinese FAQ dataset, and further apply and contextualize it to a large-scale question-matching task.
arXiv Detail & Related papers (2020-10-27T05:03:34Z) - Open Question Answering over Tables and Text [55.8412170633547]
In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Most open QA systems have considered only retrieving information from unstructured text.
We present a new large-scale dataset Open Table-and-Text Question Answering (OTT-QA) to evaluate performance on this task.
arXiv Detail & Related papers (2020-10-20T16:48:14Z) - KQA Pro: A Dataset with Explicit Compositional Programs for Complex
Question Answering over Knowledge Base [67.87878113432723]
We introduce KQA Pro, a dataset for Complex KBQA including 120K diverse natural language questions.
For each question, we provide the corresponding KoPL program and SPARQL query, so that KQA Pro serves for both KBQA and semantic parsing tasks.
arXiv Detail & Related papers (2020-07-08T03:28:04Z) - Differentiable Reasoning over a Virtual Knowledge Base [156.94984221342716]
We consider the task of answering complex multi-hop questions using a corpus as a virtual knowledge base (KB)
In particular, we describe a neural module, DrKIT, that traverses textual data like a KB, softly following paths of relations between mentions of entities in the corpus.
DrKIT is very efficient, processing 10-100x more queries per second than existing multi-hop systems.
arXiv Detail & Related papers (2020-02-25T03:13:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.