GenDec: A robust generative Question-decomposition method for Multi-hop
reasoning
- URL: http://arxiv.org/abs/2402.11166v1
- Date: Sat, 17 Feb 2024 02:21:44 GMT
- Title: GenDec: A robust generative Question-decomposition method for Multi-hop
reasoning
- Authors: Jian Wu and Linyi Yang and Yuliang Ji and Wenhao Huang and B\"orje F.
Karlsson and Manabu Okumura
- Abstract summary: Multi-hop QA involves step-by-step reasoning to answer complex questions.
Existing large language models'(LLMs) reasoning ability in multi-hop question answering remains exploration.
It is unclear whether LLMs follow a desired reasoning chain to reach the right final answer.
- Score: 32.12904215053187
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-hop QA (MHQA) involves step-by-step reasoning to answer complex
questions and find multiple relevant supporting facts. However, Existing large
language models'(LLMs) reasoning ability in multi-hop question answering
remains exploration, which is inadequate in answering multi-hop questions.
Moreover, it is unclear whether LLMs follow a desired reasoning chain to reach
the right final answer. In this paper, we propose a \textbf{gen}erative
question \textbf{dec}omposition method (GenDec) from the perspective of
explainable QA by generating independent and complete sub-questions based on
incorporating additional extracted evidence for enhancing LLMs' reasoning
ability in RAG. To demonstrate the impact, generalization, and robustness of
Gendec, we conduct two experiments, the first is combining GenDec with small QA
systems on paragraph retrieval and QA tasks. We secondly examine the reasoning
capabilities of various state-of-the-art LLMs including GPT-4 and GPT-3.5
combined with GenDec. We experiment on the HotpotQA, 2WikihopMultiHopQA,
MuSiQue, and PokeMQA datasets.
Related papers
- GenSco: Can Question Decomposition based Passage Alignment improve Question Answering? [1.5776201492893507]
"GenSco" is a novel approach of selecting passages based on the predicted decomposition of the multi-hop questions.
We evaluate on three broadly established multi-hop question answering datasets.
arXiv Detail & Related papers (2024-07-14T15:25:08Z) - Improving Question Generation with Multi-level Content Planning [70.37285816596527]
This paper addresses the problem of generating questions from a given context and an answer, specifically focusing on questions that require multi-hop reasoning across an extended context.
We propose MultiFactor, a novel QG framework based on multi-level content planning. Specifically, MultiFactor includes two components: FA-model, which simultaneously selects key phrases and generates full answers, and Q-model which takes the generated full answer as an additional input to generate questions.
arXiv Detail & Related papers (2023-10-20T13:57:01Z) - Understanding and Improving Zero-shot Multi-hop Reasoning in Generative
Question Answering [85.79940770146557]
We decompose multi-hop questions into multiple corresponding single-hop questions.
We find marked inconsistency in QA models' answers on these pairs of ostensibly identical question chains.
When trained only on single-hop questions, models generalize poorly to multi-hop questions.
arXiv Detail & Related papers (2022-10-09T11:48:07Z) - Interpretable AMR-Based Question Decomposition for Multi-hop Question
Answering [12.35571328854374]
We propose a Question Decomposition method based on Abstract Meaning Representation (QDAMR) for multi-hop QA.
We decompose a multi-hop question into simpler sub-questions and answer them in order.
Experimental results on HotpotQA demonstrate that our approach is competitive for interpretable reasoning.
arXiv Detail & Related papers (2022-06-16T23:46:33Z) - Modeling Multi-hop Question Answering as Single Sequence Prediction [88.72621430714985]
We propose a simple generative approach (PathFid) that extends the task beyond just answer generation.
PathFid explicitly models the reasoning process to resolve the answer for multi-hop questions.
Our experiments demonstrate that PathFid leads to strong performance gains on two multi-hop QA datasets.
arXiv Detail & Related papers (2022-05-18T21:57:59Z) - Calibrating Trust of Multi-Hop Question Answering Systems with
Decompositional Probes [14.302797773412543]
Multi-hop Question Answering (QA) is a challenging task since it requires an accurate aggregation of information from multiple context paragraphs.
Recent work in multi-hop QA has shown that performance can be boosted by first decomposing the questions into simpler, single-hop questions.
We show that decomposition is an effective form of probing QA systems as well as a promising approach to explanation generation.
arXiv Detail & Related papers (2022-04-16T01:03:36Z) - Ask to Understand: Question Generation for Multi-hop Question Answering [11.626390908264872]
Multi-hop Question Answering (QA) requires the machine to answer complex questions by finding scattering clues and reasoning from multiple documents.
We propose a novel method to complete multi-hop QA from the perspective of Question Generation (QG)
arXiv Detail & Related papers (2022-03-17T04:02:29Z) - QA4QG: Using Question Answering to Constrain Multi-Hop Question
Generation [54.136509061542775]
Multi-hop question generation (MQG) aims to generate complex questions which require reasoning over multiple pieces of information of the input passage.
We propose a novel framework, QA4QG, a QA-augmented BART-based framework for MQG.
Our results on the HotpotQA dataset show that QA4QG outperforms all state-of-the-art models.
arXiv Detail & Related papers (2022-02-14T08:16:47Z) - Multi-hop Question Generation with Graph Convolutional Network [58.31752179830959]
Multi-hop Question Generation (QG) aims to generate answer-related questions by aggregating and reasoning over multiple scattered evidence from different paragraphs.
We propose Multi-Hop volution Fusion Network for Question Generation (MulQG), which does context encoding in multiple hops.
Our proposed model is able to generate fluent questions with high completeness and outperforms the strongest baseline by 20.8% in the multi-hop evaluation.
arXiv Detail & Related papers (2020-10-19T06:15:36Z) - Reinforced Multi-task Approach for Multi-hop Question Generation [47.15108724294234]
We take up Multi-hop question generation, which aims at generating relevant questions based on supporting facts in the context.
We employ multitask learning with the auxiliary task of answer-aware supporting fact prediction to guide the question generator.
We demonstrate the effectiveness of our approach through experiments on the multi-hop question answering dataset, HotPotQA.
arXiv Detail & Related papers (2020-04-05T10:16:59Z) - Unsupervised Question Decomposition for Question Answering [102.56966847404287]
We propose an algorithm for One-to-N Unsupervised Sequence Sequence (ONUS) that learns to map one hard, multi-hop question to many simpler, single-hop sub-questions.
We show large QA improvements on HotpotQA over a strong baseline on the original, out-of-domain, and multi-hop dev sets.
arXiv Detail & Related papers (2020-02-22T19:40:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.