Localize, Retrieve and Fuse: A Generalized Framework for Free-Form
Question Answering over Tables
- URL: http://arxiv.org/abs/2309.11049v2
- Date: Thu, 21 Sep 2023 10:57:08 GMT
- Title: Localize, Retrieve and Fuse: A Generalized Framework for Free-Form
Question Answering over Tables
- Authors: Wenting Zhao, Ye Liu, Yao Wan, Yibo Wang, Zhongfen Deng, and Philip S.
Yu
- Abstract summary: TableQA aims at generating answers to questions grounded on a provided table.
Table-to- Graph conversion, cell localizing, external knowledge retrieval, and the fusion of table and text are proposed.
Experiments showcase the superior capabilities of TAG-QA in generating sentences that are both faithful and coherent.
- Score: 46.039687237878105
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Question answering on tabular data (a.k.a TableQA), which aims at generating
answers to questions grounded on a provided table, has gained significant
attention recently. Prior work primarily produces concise factual responses
through information extraction from individual or limited table cells, lacking
the ability to reason across diverse table cells. Yet, the realm of free-form
TableQA, which demands intricate strategies for selecting relevant table cells
and the sophisticated integration and inference of discrete data fragments,
remains mostly unexplored. To this end, this paper proposes a generalized
three-stage approach: Table-to- Graph conversion and cell localizing, external
knowledge retrieval, and the fusion of table and text (called TAG-QA), to
address the challenge of inferring long free-form answers in generative
TableQA. In particular, TAG-QA (1) locates relevant table cells using a graph
neural network to gather intersecting cells between relevant rows and columns,
(2) leverages external knowledge from Wikipedia, and (3) generates answers by
integrating both tabular data and natural linguistic information. Experiments
showcase the superior capabilities of TAG-QA in generating sentences that are
both faithful and coherent, particularly when compared to several
state-of-the-art baselines. Notably, TAG-QA surpasses the robust pipeline-based
baseline TAPAS by 17% and 14% in terms of BLEU-4 and PARENT F-score,
respectively. Furthermore, TAG-QA outperforms the end-to-end model T5 by 16%
and 12% on BLEU-4 and PARENT F-score, respectively.
Related papers
- Knowledge in Triples for LLMs: Enhancing Table QA Accuracy with Semantic Extraction [1.0968343822308813]
This paper proposes a novel approach that extracts triples straightforward from tabular data and integrates it with a retrieval-augmented generation (RAG) model to enhance the accuracy, coherence, and contextual richness of responses generated by a fine-tuned GPT-3.5-turbo-0125 model.
Our approach significantly outperforms existing baselines on the FeTaQA dataset, particularly excelling in Sacre-BLEU and ROUGE metrics.
arXiv Detail & Related papers (2024-09-21T16:46:15Z) - KET-QA: A Dataset for Knowledge Enhanced Table Question Answering [63.56707527868466]
We propose to use a knowledge base (KB) as the external knowledge source for TableQA.
Every question requires the integration of information from both the table and the sub-graph to be answered.
We design a retriever-reasoner structured pipeline model to extract pertinent information from the vast knowledge sub-graph.
arXiv Detail & Related papers (2024-05-13T18:26:32Z) - TACR: A Table-alignment-based Cell-selection and Reasoning Model for
Hybrid Question-Answering [31.79113994947629]
We propose a Table-alignment-based Cell-selection and Reasoning model (TACR) for hybrid text and table QA.
In evidence retrieval, we design a table-question-alignment enhanced cell-selection method to retrieve fine-grained evidence.
In answer reasoning, we incorporate a QA module that treats the row containing selected cells as context.
arXiv Detail & Related papers (2023-05-24T03:42:44Z) - MultiTabQA: Generating Tabular Answers for Multi-Table Question
Answering [61.48881995121938]
Real-world queries are complex in nature, often over multiple tables in a relational database or web page.
Our model, MultiTabQA, not only answers questions over multiple tables, but also generalizes to generate tabular answers.
arXiv Detail & Related papers (2023-05-22T08:25:15Z) - Multi-Row, Multi-Span Distant Supervision For Table+Text Question [33.809732338627136]
Question answering (QA) over tables and linked text, also called TextTableQA, has witnessed significant research in recent years.
We present MITQA, a transformer-based TextTableQA system that is explicitly designed to cope with distant supervision along both these axes.
arXiv Detail & Related papers (2021-12-14T12:48:19Z) - TAT-QA: A Question Answering Benchmark on a Hybrid of Tabular and
Textual Content in Finance [71.76018597965378]
We build a new large-scale Question Answering dataset containing both Tabular And Textual data, named TAT-QA.
We propose a novel QA model termed TAGOP, which is capable of reasoning over both tables and text.
arXiv Detail & Related papers (2021-05-17T06:12:06Z) - FeTaQA: Free-form Table Question Answering [33.018256483762386]
We introduce FeTaQA, a new dataset with 10K Wikipedia-based table, question, free-form answer, supporting table cells pairs.
FeTaQA yields a more challenging table question answering setting because it requires generating free-form text answers after retrieval, inference, and integration of multiple discontinuous facts from a structured knowledge source.
arXiv Detail & Related papers (2021-04-01T09:59:40Z) - TCN: Table Convolutional Network for Web Table Interpretation [52.32515851633981]
We propose a novel table representation learning approach considering both the intra- and inter-table contextual information.
Our method can outperform competitive baselines by +4.8% of F1 for column type prediction and by +4.1% of F1 for column pairwise relation prediction.
arXiv Detail & Related papers (2021-02-17T02:18:10Z) - Open Question Answering over Tables and Text [55.8412170633547]
In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Most open QA systems have considered only retrieving information from unstructured text.
We present a new large-scale dataset Open Table-and-Text Question Answering (OTT-QA) to evaluate performance on this task.
arXiv Detail & Related papers (2020-10-20T16:48:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.