MuGER$^2$: Multi-Granularity Evidence Retrieval and Reasoning for Hybrid
Question Answering
- URL: http://arxiv.org/abs/2210.10350v1
- Date: Wed, 19 Oct 2022 07:36:03 GMT
- Title: MuGER$^2$: Multi-Granularity Evidence Retrieval and Reasoning for Hybrid
Question Answering
- Authors: Yingyao Wang, Junwei Bao, Chaoqun Duan, Youzheng Wu, Xiaodong He and
Tiejun Zhao
- Abstract summary: Hybrid question answering (HQA) aims to answer questions over heterogeneous data, including tables and passages linked to table cells.
We propose MuGER$2$, a Multi-Granularity Evidence Retrieval and Reasoning approach.
Experiment results on the HybridQA dataset show that MuGER$2$ significantly boosts the HQA performance.
- Score: 32.850210766905505
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Hybrid question answering (HQA) aims to answer questions over heterogeneous
data, including tables and passages linked to table cells. The heterogeneous
data can provide different granularity evidence to HQA models, e.t., column,
row, cell, and link. Conventional HQA models usually retrieve coarse- or
fine-grained evidence to reason the answer. Through comparison, we find that
coarse-grained evidence is easier to retrieve but contributes less to the
reasoner, while fine-grained evidence is the opposite. To preserve the
advantage and eliminate the disadvantage of different granularity evidence, we
propose MuGER$^2$, a Multi-Granularity Evidence Retrieval and Reasoning
approach. In evidence retrieval, a unified retriever is designed to learn the
multi-granularity evidence from the heterogeneous data. In answer reasoning, an
evidence selector is proposed to navigate the fine-grained evidence for the
answer reader based on the learned multi-granularity evidence. Experiment
results on the HybridQA dataset show that MuGER$^2$ significantly boosts the
HQA performance. Further ablation analysis verifies the effectiveness of both
the retrieval and reasoning designs.
Related papers
- Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering [41.990482015732574]
We propose a novel evidence-enhanced triplet generation framework, EATQA, to predict all the combinations of (Question, Evidence, Answer) triplet.
We bridge the distribution gap to distill the knowledge from evidence in inference stage.
Our framework ensures the model to learn the logical relation between query, evidence and answer, which simultaneously improves the evidence generation and query answering.
arXiv Detail & Related papers (2024-08-27T13:07:07Z) - Chain-of-Discussion: A Multi-Model Framework for Complex Evidence-Based Question Answering [55.295699268654545]
We propose a novel Chain-of-Discussion framework to leverage the synergy among open-source Large Language Models.
Our experiments show that discussions among multiple LLMs play a vital role in enhancing the quality of answers.
arXiv Detail & Related papers (2024-02-26T05:31:34Z) - Multi-grained Evidence Inference for Multi-choice Reading Comprehension [62.0773160298008]
Multi-choice Machine Reading (MRC) is a major and challenging task for machines to answer questions according to provided options.
We propose a novel general-purpose model enhancement which integrates multi-grained evidence comprehensively, named Multi-grained evidence inferencer (Mugen)
Mugen extracts three different granularities of evidence, and integrates evidence with the original passages, achieving significant and consistent performance improvement on four multi-choice MRC benchmarks.
arXiv Detail & Related papers (2023-10-27T11:36:18Z) - An Empirical Comparison of LM-based Question and Answer Generation
Methods [79.31199020420827]
Question and answer generation (QAG) consists of generating a set of question-answer pairs given a context.
In this paper, we establish baselines with three different QAG methodologies that leverage sequence-to-sequence language model (LM) fine-tuning.
Experiments show that an end-to-end QAG model, which is computationally light at both training and inference times, is generally robust and outperforms other more convoluted approaches.
arXiv Detail & Related papers (2023-05-26T14:59:53Z) - TACR: A Table-alignment-based Cell-selection and Reasoning Model for
Hybrid Question-Answering [31.79113994947629]
We propose a Table-alignment-based Cell-selection and Reasoning model (TACR) for hybrid text and table QA.
In evidence retrieval, we design a table-question-alignment enhanced cell-selection method to retrieve fine-grained evidence.
In answer reasoning, we incorporate a QA module that treats the row containing selected cells as context.
arXiv Detail & Related papers (2023-05-24T03:42:44Z) - ReasonChainQA: Text-based Complex Question Answering with Explainable
Evidence Chains [15.837457557803507]
We present a benchmark textbfReasonChainQA with explanatory and explicit evidence chains.
ReasonChainQA consists of two subtasks: answer generation and evidence chains extraction, it also contains higher diversity for multi-hop questions.
Additional experiment on supervised and unsupervised retrieval fully indicates the significance of ReasonChainQA.
arXiv Detail & Related papers (2022-10-17T06:07:39Z) - Grow-and-Clip: Informative-yet-Concise Evidence Distillation for Answer
Explanation [22.20733260041759]
We argue that the evidences of an answer is critical to enhancing the interpretability of QA models.
We are the first to explicitly define the concept of evidence as the supporting facts in a context which are informative, concise, and readable.
We propose Grow-and-Clip Evidence Distillation (GCED) algorithm to extract evidences from the contexts by trade-off informativeness, conciseness, and readability.
arXiv Detail & Related papers (2022-01-13T17:18:17Z) - Distantly-Supervised Evidence Retrieval Enables Question Answering
without Evidence Annotation [19.551623461082617]
Open-domain question answering answers a question based on evidence retrieved from a large corpus.
This paper investigates whether models can learn to find evidence from a large corpus, with only distant supervision from answer labels for model training.
We introduce a novel approach (DistDR) that iteratively improves over a weak retriever by alternately finding evidence from the up-to-date model and encouraging the model to learn the most likely evidence.
arXiv Detail & Related papers (2021-10-10T20:01:27Z) - REM-Net: Recursive Erasure Memory Network for Commonsense Evidence
Refinement [130.8875535449478]
REM-Net is equipped with a module to refine the evidence by erasing the low-quality evidence that does not explain the question answering.
Instead of retrieving evidence from existing knowledge bases, REM-Net leverages a pre-trained generative model to generate candidate evidence customized for the question.
The results demonstrate the performance of REM-Net and show that the refined evidence is explainable.
arXiv Detail & Related papers (2020-12-24T10:07:32Z) - Leveraging Passage Retrieval with Generative Models for Open Domain
Question Answering [61.394478670089065]
Generative models for open domain question answering have proven to be competitive, without resorting to external knowledge.
We investigate how much these models can benefit from retrieving text passages, potentially containing evidence.
We observe that the performance of this method significantly improves when increasing the number of retrieved passages.
arXiv Detail & Related papers (2020-07-02T17:44:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.