Answering Numerical Reasoning Questions in Table-Text Hybrid Contents
with Graph-based Encoder and Tree-based Decoder
- URL: http://arxiv.org/abs/2209.07692v1
- Date: Fri, 16 Sep 2022 03:15:12 GMT
- Title: Answering Numerical Reasoning Questions in Table-Text Hybrid Contents
with Graph-based Encoder and Tree-based Decoder
- Authors: Fangyu Lei, Shizhu He, Xiang Li, Jun Zhao, Kang Liu
- Abstract summary: This paper proposes a textbfRelational textbfGraph enhanced textbfHybrid table-text textbfNumerical reasoning model with textbfRegHNT.
It models the numerical question answering over table-text hybrid contents as an expression tree generation task. We validated our model on the publicly available table-text hybrid QA benchmark (TAT-QA)
- Score: 19.429216786198577
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the real-world question answering scenarios, hybrid form combining both
tabular and textual contents has attracted more and more attention, among which
numerical reasoning problem is one of the most typical and challenging
problems. Existing methods usually adopt encoder-decoder framework to represent
hybrid contents and generate answers. However, it can not capture the rich
relationship among numerical value, table schema, and text information on the
encoder side. The decoder uses a simple predefined operator classifier which is
not flexible enough to handle numerical reasoning processes with diverse
expressions. To address these problems, this paper proposes a
\textbf{Re}lational \textbf{G}raph enhanced \textbf{H}ybrid table-text
\textbf{N}umerical reasoning model with \textbf{T}ree decoder
(\textbf{RegHNT}). It models the numerical question answering over table-text
hybrid contents as an expression tree generation task. Moreover, we propose a
novel relational graph modeling method, which models alignment between
questions, tables, and paragraphs. We validated our model on the publicly
available table-text hybrid QA benchmark (TAT-QA). The proposed RegHNT
significantly outperform the baseline model and achieve state-of-the-art
results\footnote{We openly released the source code and data
at~\url{https://github.com/lfy79001/RegHNT}}~(2022-05-05).
Related papers
- Relation Rectification in Diffusion Model [64.84686527988809]
We introduce a novel task termed Relation Rectification, aiming to refine the model to accurately represent a given relationship it initially fails to generate.
We propose an innovative solution utilizing a Heterogeneous Graph Convolutional Network (HGCN)
The lightweight HGCN adjusts the text embeddings generated by the text encoder, ensuring the accurate reflection of the textual relation in the embedding space.
arXiv Detail & Related papers (2024-03-29T15:54:36Z) - An Autoregressive Text-to-Graph Framework for Joint Entity and Relation
Extraction [4.194768796374315]
We propose a novel method for joint entity and relation extraction from unstructured text by framing it as a conditional sequence generation problem.
It generates a linearized graph where nodes represent text spans and edges represent relation triplets.
Our method employs a transformer encoder-decoder architecture with pointing mechanism on a dynamic vocabulary of spans and relation types.
arXiv Detail & Related papers (2024-01-02T18:32:14Z) - Answer Candidate Type Selection: Text-to-Text Language Model for Closed
Book Question Answering Meets Knowledge Graphs [62.20354845651949]
We present a novel approach which works on top of the pre-trained Text-to-Text QA system to address this issue.
Our simple yet effective method performs filtering and re-ranking of generated candidates based on their types derived from Wikidata "instance_of" property.
arXiv Detail & Related papers (2023-10-10T20:49:43Z) - Multi-View Graph Representation Learning for Answering Hybrid Numerical
Reasoning Question [13.321467396155116]
The paper proposes a Multi-View Graph (MVG) to take the relations among the granularity into account and capture the relations from multiple view.
We validate our model on the publicly available table-text hybrid QA benchmark (TAT-QA) and outperform the state-of-the-art model.
arXiv Detail & Related papers (2023-05-05T12:00:58Z) - Mixed-modality Representation Learning and Pre-training for Joint
Table-and-Text Retrieval in OpenQA [85.17249272519626]
An optimized OpenQA Table-Text Retriever (OTTeR) is proposed.
We conduct retrieval-centric mixed-modality synthetic pre-training.
OTTeR substantially improves the performance of table-and-text retrieval on the OTT-QA dataset.
arXiv Detail & Related papers (2022-10-11T07:04:39Z) - Text-to-Table: A New Way of Information Extraction [8.326657025342042]
We study a new problem setting of information extraction (IE), referred to as text-to-table.
In text-to-table, given a text, one creates a table or several tables expressing the main content of the text.
We make use of four existing table-to-text datasets in our experiments on text-to-table.
arXiv Detail & Related papers (2021-09-06T19:35:46Z) - Multi-modal Retrieval of Tables and Texts Using Tri-encoder Models [2.5621280373733604]
Some questions cannot be answered by text alone but require information stored in tables.
We present an approach for retrieving both texts and tables relevant to a question by jointly encoding texts, tables and questions into a single vector space.
We release the newly created multi-modal dataset to the community so that it can be used for training and evaluation.
arXiv Detail & Related papers (2021-08-09T14:02:00Z) - TAT-QA: A Question Answering Benchmark on a Hybrid of Tabular and
Textual Content in Finance [71.76018597965378]
We build a new large-scale Question Answering dataset containing both Tabular And Textual data, named TAT-QA.
We propose a novel QA model termed TAGOP, which is capable of reasoning over both tables and text.
arXiv Detail & Related papers (2021-05-17T06:12:06Z) - Rethinking Positional Encoding in Language Pre-training [111.2320727291926]
We show that in absolute positional encoding, the addition operation applied on positional embeddings and word embeddings brings mixed correlations.
We propose a new positional encoding method called textbfTransformer with textbfUntied textPositional textbfEncoding (T)
arXiv Detail & Related papers (2020-06-28T13:11:02Z) - Semantic Graphs for Generating Deep Questions [98.5161888878238]
We propose a novel framework which first constructs a semantic-level graph for the input document and then encodes the semantic graph by introducing an attention-based GGNN (Att-GGNN)
On the HotpotQA deep-question centric dataset, our model greatly improves performance over questions requiring reasoning over multiple facts, leading to state-of-the-art performance.
arXiv Detail & Related papers (2020-04-27T10:52:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.