Fact-Tree Reasoning for N-ary Question Answering over Knowledge Graphs
- URL: http://arxiv.org/abs/2108.08297v1
- Date: Tue, 17 Aug 2021 13:27:49 GMT
- Title: Fact-Tree Reasoning for N-ary Question Answering over Knowledge Graphs
- Authors: Yao Zhang, Peiyao Li, Hongru Liang, Adam Jatowt, Zhenglu Yang
- Abstract summary: We propose a novel fact-tree reasoning framework, through transforming the question into a fact tree and performing iterative fact reasoning on it to predict the correct answer.
We demonstrate that the proposed fact-tree reasoning framework has the desired advantage of high answer prediction accuracy.
- Score: 21.87251293779023
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the question answering(QA) task, multi-hop reasoning framework has been
extensively studied in recent years to perform more efficient and interpretable
answer reasoning on the Knowledge Graph(KG). However, multi-hop reasoning is
inapplicable for answering n-ary fact questions due to its linear reasoning
nature. We discover that there are two feasible improvements: 1) upgrade the
basic reasoning unit from entity or relation to fact; and 2) upgrade the
reasoning structure from chain to tree. Based on these, we propose a novel
fact-tree reasoning framework, through transforming the question into a fact
tree and performing iterative fact reasoning on it to predict the correct
answer. Through a comprehensive evaluation on the n-ary fact KGQA dataset
introduced by this work, we demonstrate that the proposed fact-tree reasoning
framework has the desired advantage of high answer prediction accuracy. In
addition, we also evaluate the fact-tree reasoning framework on two binary KGQA
datasets and show that our approach also has a strong reasoning ability
compared with several excellent baselines. This work has direct implications
for exploring complex reasoning scenarios and provides a preliminary baseline
approach.
Related papers
- GRSQA -- Graph Reasoning-Structured Question Answering Dataset [50.223851616680754]
We introduce the Graph Reasoning-Structured Question Answering dataset (GRS-QA), which includes both semantic contexts and reasoning structures for QA pairs.
Unlike existing M-QA datasets, GRS-QA explicitly captures intricate reasoning pathways by constructing reasoning graphs.
Our empirical analysis reveals that LLMs perform differently when handling questions with varying reasoning structures.
arXiv Detail & Related papers (2024-11-01T05:14:03Z) - Probabilistic Tree-of-thought Reasoning for Answering
Knowledge-intensive Complex Questions [93.40614719648386]
Large language models (LLMs) are capable of answering knowledge-intensive complex questions with chain-of-thought (CoT) reasoning.
Recent works turn to retrieving external knowledge to augment CoT reasoning.
We propose a novel approach: Probabilistic Tree-of-thought Reasoning (ProbTree)
arXiv Detail & Related papers (2023-11-23T12:52:37Z) - Leveraging Structured Information for Explainable Multi-hop Question
Answering and Reasoning [14.219239732584368]
In this work, we investigate constructing and leveraging extracted semantic structures (graphs) for multi-hop question answering.
Empirical results and human evaluations show that our framework: generates more faithful reasoning chains and substantially improves the QA performance on two benchmark datasets.
arXiv Detail & Related papers (2023-11-07T05:32:39Z) - Reasoning over Hierarchical Question Decomposition Tree for Explainable
Question Answering [83.74210749046551]
We propose to leverage question decomposing for heterogeneous knowledge integration.
We propose a novel two-stage XQA framework, Reasoning over Hierarchical Question Decomposition Tree (RoHT)
Experiments on complex QA datasets KQA Pro and Musique show that our framework outperforms SOTA methods significantly.
arXiv Detail & Related papers (2023-05-24T11:45:59Z) - ReCEval: Evaluating Reasoning Chains via Correctness and Informativeness [67.49087159888298]
ReCEval is a framework that evaluates reasoning chains via two key properties: correctness and informativeness.
We show that ReCEval effectively identifies various error types and yields notable improvements compared to prior methods.
arXiv Detail & Related papers (2023-04-21T02:19:06Z) - MetaLogic: Logical Reasoning Explanations with Fine-Grained Structure [129.8481568648651]
We propose a benchmark to investigate models' logical reasoning capabilities in complex real-life scenarios.
Based on the multi-hop chain of reasoning, the explanation form includes three main components.
We evaluate the current best models' performance on this new explanation form.
arXiv Detail & Related papers (2022-10-22T16:01:13Z) - Dynamic Semantic Graph Construction and Reasoning for Explainable
Multi-hop Science Question Answering [50.546622625151926]
We propose a new framework to exploit more valid facts while obtaining explainability for multi-hop QA.
Our framework contains three new ideas: (a) tt AMR-SG, an AMR-based Semantic Graph, constructed by candidate fact AMRs to uncover any hop relations among question, answer and multiple facts, (b) a novel path-based fact analytics approach exploiting tt AMR-SG to extract active facts from a large fact pool to answer questions, and (c) a fact-level relation modeling leveraging graph convolution network (GCN) to guide the reasoning process.
arXiv Detail & Related papers (2021-05-25T09:14:55Z) - Explaining Answers with Entailment Trees [16.555369850015055]
We aim to explain answers by showing how evidence leads to the answer in a systematic way.
Our approach is to generate explanations in the form of entailment trees, namely a tree of entailment steps from facts that are known, through intermediate conclusions, to the final answer.
To train a model with this skill, we created ENTAILMENTBANK, the first dataset to contain multistep entailment trees.
arXiv Detail & Related papers (2021-04-17T23:13:56Z) - ExplanationLP: Abductive Reasoning for Explainable Science Question
Answering [4.726777092009554]
This paper frames question answering as an abductive reasoning problem.
We construct plausible explanations for each choice and then selecting the candidate with the best explanation as the final answer.
Our system, ExplanationLP, elicits explanations by constructing a weighted graph of relevant facts for each candidate answer.
arXiv Detail & Related papers (2020-10-25T14:49:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.