BreakingBERT@IITK at SemEval-2021 Task 9 : Statement Verification and
Evidence Finding with Tables
- URL: http://arxiv.org/abs/2104.03071v1
- Date: Wed, 7 Apr 2021 11:41:07 GMT
- Title: BreakingBERT@IITK at SemEval-2021 Task 9 : Statement Verification and
Evidence Finding with Tables
- Authors: Aditya Jindal, Ankur Gupta, Jaya Srivastava, Preeti Menghwani, Vijit
Malik, Vishesh Kaushik, Ashutosh Modi
- Abstract summary: We tackle the problem of fact verification and evidence finding over tabular data.
We make a comparison of the baselines and state-of-the-art approaches over the given SemTabFact dataset.
We also propose a novel approach CellBERT to solve evidence finding as a form of the Natural Language Inference task.
- Score: 1.78256232654567
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recently, there has been an interest in factual verification and prediction
over structured data like tables and graphs. To circumvent any false news
incident, it is necessary to not only model and predict over structured data
efficiently but also to explain those predictions. In this paper, as part of
the SemEval-2021 Task 9, we tackle the problem of fact verification and
evidence finding over tabular data. There are two subtasks. Given a table and a
statement/fact, subtask A determines whether the statement is inferred from the
tabular data, and subtask B determines which cells in the table provide
evidence for the former subtask. We make a comparison of the baselines and
state-of-the-art approaches over the given SemTabFact dataset. We also propose
a novel approach CellBERT to solve evidence finding as a form of the Natural
Language Inference task. We obtain a 3-way F1 score of 0.69 on subtask A and an
F1 score of 0.65 on subtask B.
Related papers
- Chain-of-Table: Evolving Tables in the Reasoning Chain for Table
Understanding [79.9461269253121]
We propose the Chain-of-Table framework, where tabular data is explicitly used in the reasoning chain as a proxy for intermediate thoughts.
Chain-of-Table achieves new state-of-the-art performance on WikiTQ, FeTaQA, and TabFact benchmarks.
arXiv Detail & Related papers (2024-01-09T07:46:26Z) - QTSumm: Query-Focused Summarization over Tabular Data [58.62152746690958]
People primarily consult tables to conduct data analysis or answer specific questions.
We define a new query-focused table summarization task, where text generation models have to perform human-like reasoning.
We introduce a new benchmark named QTSumm for this task, which contains 7,111 human-annotated query-summary pairs over 2,934 tables.
arXiv Detail & Related papers (2023-05-23T17:43:51Z) - Graph Reasoning with Context-Aware Linearization for Interpretable Fact
Extraction and Verification [18.80709296426521]
This paper presents an end-to-end system for fact extraction and verification using textual and tabular evidence.
Our best performing system achieves a FEVEROUS score of 0.23 and 53% label accuracy on the blind test data.
arXiv Detail & Related papers (2021-09-25T12:05:28Z) - Zero-Shot Information Extraction as a Unified Text-to-Triple Translation [56.01830747416606]
We cast a suite of information extraction tasks into a text-to-triple translation framework.
We formalize the task as a translation between task-specific input text and output triples.
We study the zero-shot performance of this framework on open information extraction.
arXiv Detail & Related papers (2021-09-23T06:54:19Z) - Volta at SemEval-2021 Task 9: Statement Verification and Evidence
Finding with Tables using TAPAS and Transfer Learning [19.286478269708592]
We present our systems to solve Task 9 of SemEval-2021: Statement Verification and Evidence Finding with Tables.
The task consists of two subtasks: (A) Given a table and a statement, predicting whether the table supports the statement and (B) Predicting which cells in the table provide evidence for/against the statement.
Our systems achieve an F1 score of 67.34 in subtask A three-way classification, 72.89 in subtask A two-way classification, and 62.95 in subtask B.
arXiv Detail & Related papers (2021-06-01T06:06:29Z) - SemEval-2021 Task 9: Fact Verification and Evidence Finding for Tabular
Data in Scientific Documents (SEM-TAB-FACTS) [0.0]
SEM-TAB-FACTS featured two sub-tasks.
In sub-task A, the goal was to determine if a statement is supported, refuted or unknown in relation to a table.
In sub-task B, the focus was on identifying the specific cells of a table that provide evidence for the statement.
arXiv Detail & Related papers (2021-05-28T17:21:11Z) - TAT-QA: A Question Answering Benchmark on a Hybrid of Tabular and
Textual Content in Finance [71.76018597965378]
We build a new large-scale Question Answering dataset containing both Tabular And Textual data, named TAT-QA.
We propose a novel QA model termed TAGOP, which is capable of reasoning over both tables and text.
arXiv Detail & Related papers (2021-05-17T06:12:06Z) - TABBIE: Pretrained Representations of Tabular Data [22.444607481407633]
We devise a simple pretraining objective that learns exclusively from tabular data.
Unlike competing approaches, our model (TABBIE) provides embeddings of all table substructures.
A qualitative analysis of our model's learned cell, column, and row representations shows that it understands complex table semantics and numerical trends.
arXiv Detail & Related papers (2021-05-06T11:15:16Z) - Sattiy at SemEval-2021 Task 9: An Ensemble Solution for Statement
Verification and Evidence Finding with Tables [4.691435917434472]
This paper describes sattiy team's system in SemEval-2021 task 9: Statement Verification and Evidence Finding with Tables (SEM-TAB-FACT)
This competition aims to verify statements and to find evidence from tables for scientific articles.
In this paper, we exploited ensemble models of pre-trained language models over tables, TaPas and TaBERT, for Task A and adjust the result based on some rules extracted for Task B.
arXiv Detail & Related papers (2021-04-21T06:11:49Z) - TaBERT: Pretraining for Joint Understanding of Textual and Tabular Data [113.29476656550342]
We present TaBERT, a pretrained LM that jointly learns representations for NL sentences and tables.
TaBERT is trained on a large corpus of 26 million tables and their English contexts.
Implementation of the model will be available at http://fburl.com/TaBERT.
arXiv Detail & Related papers (2020-05-17T17:26:40Z) - ToTTo: A Controlled Table-To-Text Generation Dataset [61.83159452483026]
ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples.
We introduce a dataset construction process where annotators directly revise existing candidate sentences from Wikipedia.
While usually fluent, existing methods often hallucinate phrases that are not supported by the table.
arXiv Detail & Related papers (2020-04-29T17:53:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.