Open Question Answering over Tables and Text
- URL: http://arxiv.org/abs/2010.10439v2
- Date: Wed, 10 Feb 2021 08:21:18 GMT
- Title: Open Question Answering over Tables and Text
- Authors: Wenhu Chen, Ming-Wei Chang, Eva Schlinger, William Wang, William W.
Cohen
- Abstract summary: In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Most open QA systems have considered only retrieving information from unstructured text.
We present a new large-scale dataset Open Table-and-Text Question Answering (OTT-QA) to evaluate performance on this task.
- Score: 55.8412170633547
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In open question answering (QA), the answer to a question is produced by
retrieving and then analyzing documents that might contain answers to the
question. Most open QA systems have considered only retrieving information from
unstructured text. Here we consider for the first time open QA over both
tabular and textual data and present a new large-scale dataset Open
Table-and-Text Question Answering (OTT-QA) to evaluate performance on this
task. Most questions in OTT-QA require multi-hop inference across tabular data
and unstructured text, and the evidence required to answer a question can be
distributed in different ways over these two types of input, making evidence
retrieval challenging -- our baseline model using an iterative retriever and
BERT-based reader achieves an exact match score less than 10%. We then propose
two novel techniques to address the challenge of retrieving and aggregating
evidence for OTT-QA. The first technique is to use "early fusion" to group
multiple highly relevant tabular and textual units into a fused block, which
provides more context for the retriever to search for. The second technique is
to use a cross-block reader to model the cross-dependency between multiple
retrieved evidence with global-local sparse attention. Combining these two
techniques improves the score significantly, to above 27%.
Related papers
- TTQA-RS- A break-down prompting approach for Multi-hop Table-Text Question Answering with Reasoning and Summarization [3.531533402602335]
Multi-hop table-text QA requires multiple hops between the table and text.
Our model uses an enhanced retriever for table-text information retrieval.
Our experiments demonstrate the potential of prompt-based approaches.
arXiv Detail & Related papers (2024-06-20T20:55:38Z) - MFORT-QA: Multi-hop Few-shot Open Rich Table Question Answering [3.1651118728570635]
In today's fast-paced industry, professionals face the challenge of summarizing a large number of documents and extracting vital information from them on a daily basis.
To address this challenge, the approach of Table Question Answering (QA) has been developed to extract the relevant information.
Recent advancements in Large Language Models (LLMs) have opened up new possibilities for extracting information from tabular data using prompts.
arXiv Detail & Related papers (2024-03-28T03:14:18Z) - UniKGQA: Unified Retrieval and Reasoning for Solving Multi-hop Question
Answering Over Knowledge Graph [89.98762327725112]
Multi-hop Question Answering over Knowledge Graph(KGQA) aims to find the answer entities that are multiple hops away from the topic entities mentioned in a natural language question.
We propose UniKGQA, a novel approach for multi-hop KGQA task, by unifying retrieval and reasoning in both model architecture and parameter learning.
arXiv Detail & Related papers (2022-12-02T04:08:09Z) - Mixed-modality Representation Learning and Pre-training for Joint
Table-and-Text Retrieval in OpenQA [85.17249272519626]
An optimized OpenQA Table-Text Retriever (OTTeR) is proposed.
We conduct retrieval-centric mixed-modality synthetic pre-training.
OTTeR substantially improves the performance of table-and-text retrieval on the OTT-QA dataset.
arXiv Detail & Related papers (2022-10-11T07:04:39Z) - Multifaceted Improvements for Conversational Open-Domain Question
Answering [54.913313912927045]
We propose a framework with Multifaceted Improvements for Conversational open-domain Question Answering (MICQA)
Firstly, the proposed KL-divergence based regularization is able to lead to a better question understanding for retrieval and answer reading.
Second, the added post-ranker module can push more relevant passages to the top placements and be selected for reader with a two-aspect constrains.
Third, the well designed curriculum learning strategy effectively narrows the gap between the golden passage settings of training and inference, and encourages the reader to find true answer without the golden passage assistance.
arXiv Detail & Related papers (2022-04-01T07:54:27Z) - When Retriever-Reader Meets Scenario-Based Multiple-Choice Questions [15.528174963480614]
We propose a joint retriever-reader model called QAVES where the retriever is implicitly supervised only using relevance labels via a novel word weighting mechanism.
QAVES significantly outperforms a variety of strong baselines on multiple-choice questions in three SQA datasets.
arXiv Detail & Related papers (2021-08-31T14:32:04Z) - FeTaQA: Free-form Table Question Answering [33.018256483762386]
We introduce FeTaQA, a new dataset with 10K Wikipedia-based table, question, free-form answer, supporting table cells pairs.
FeTaQA yields a more challenging table question answering setting because it requires generating free-form text answers after retrieval, inference, and integration of multiple discontinuous facts from a structured knowledge source.
arXiv Detail & Related papers (2021-04-01T09:59:40Z) - Tradeoffs in Sentence Selection Techniques for Open-Domain Question
Answering [54.541952928070344]
We describe two groups of models for sentence selection: QA-based approaches, which run a full-fledged QA system to identify answer candidates, and retrieval-based models, which find parts of each passage specifically related to each question.
We show that very lightweight QA models can do well at this task, but retrieval-based models are faster still.
arXiv Detail & Related papers (2020-09-18T23:39:15Z) - Answering Any-hop Open-domain Questions with Iterative Document
Reranking [62.76025579681472]
We propose a unified QA framework to answer any-hop open-domain questions.
Our method consistently achieves performance comparable to or better than the state-of-the-art on both single-hop and multi-hop open-domain QA datasets.
arXiv Detail & Related papers (2020-09-16T04:31:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.