TabPert: An Effective Platform for Tabular Perturbation
- URL: http://arxiv.org/abs/2108.00603v1
- Date: Mon, 2 Aug 2021 02:37:48 GMT
- Title: TabPert: An Effective Platform for Tabular Perturbation
- Authors: Nupur Jain, Vivek Gupta, Anshul Rai, Gaurav Kumar
- Abstract summary: TabPert allows a user to update a table, change its associated hypotheses, change their labels, and highlight rows that are important for hypothesis classification.
These counterfactual tables and hypotheses, as well as the metadata, can then be used to explore an existing model's shortcomings methodically and quantitatively.
- Score: 6.555691728969102
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To truly grasp reasoning ability, a Natural Language Inference model should
be evaluated on counterfactual data. TabPert facilitates this by assisting in
the generation of such counterfactual data for assessing model tabular
reasoning issues. TabPert allows a user to update a table, change its
associated hypotheses, change their labels, and highlight rows that are
important for hypothesis classification. TabPert also captures information
about the techniques used to automatically produce the table, as well as the
strategies employed to generate the challenging hypotheses. These
counterfactual tables and hypotheses, as well as the metadata, can then be used
to explore an existing model's shortcomings methodically and quantitatively.
Related papers
- A Simple Baseline for Predicting Events with Auto-Regressive Tabular Transformers [70.20477771578824]
Existing approaches to event prediction include time-aware positional embeddings, learned row and field encodings, and oversampling methods for addressing class imbalance.
We propose a simple but flexible baseline using standard autoregressive LLM-style transformers with elementary positional embeddings and a causal language modeling objective.
Our baseline outperforms existing approaches across popular datasets and can be employed for various use-cases.
arXiv Detail & Related papers (2024-10-14T15:59:16Z) - FLEXTAF: Enhancing Table Reasoning with Flexible Tabular Formats [48.47559543509975]
We propose FLEXTAF-Single and FLEXTAF-Vote to enhance table reasoning performance by employing flexible formats.
Our experiments on WikiTableQuestions and TabFact reveal significant improvements, with average gains of 2.3% and 4.8%.
arXiv Detail & Related papers (2024-08-16T17:00:11Z) - AdapTable: Test-Time Adaptation for Tabular Data via Shift-Aware Uncertainty Calibrator and Label Distribution Handler [29.395855812763617]
We propose AdapTable, a framework for adapting machine learning models to target data without accessing source data.
AdapTable operates in two stages: 1) calibrating model predictions using a shift-aware uncertainty calibrator, and 2) adjusting these predictions to match the target label distribution with a label distribution handler.
Our results demonstrate AdapTable's ability to handle various real-world distribution shifts, achieving up to a 16% improvement on the dataset.
arXiv Detail & Related papers (2024-07-15T15:02:53Z) - LaTable: Towards Large Tabular Models [63.995130144110156]
Tabular generative foundation models are hard to build due to the heterogeneous feature spaces of different datasets.
LaTable is a novel diffusion model that addresses these challenges and can be trained across different datasets.
We find that LaTable outperforms baselines on in-distribution generation, and that finetuning LaTable can generate out-of-distribution datasets better with fewer samples.
arXiv Detail & Related papers (2024-06-25T16:03:50Z) - Chain-of-Table: Evolving Tables in the Reasoning Chain for Table
Understanding [79.9461269253121]
We propose the Chain-of-Table framework, where tabular data is explicitly used in the reasoning chain as a proxy for intermediate thoughts.
Chain-of-Table achieves new state-of-the-art performance on WikiTQ, FeTaQA, and TabFact benchmarks.
arXiv Detail & Related papers (2024-01-09T07:46:26Z) - TabRepo: A Large Scale Repository of Tabular Model Evaluations and its AutoML Applications [9.457938949410583]
TabRepo is a new dataset of model evaluations and predictions.
It contains the predictions and metrics of 1310 models evaluated on 200 datasets.
arXiv Detail & Related papers (2023-11-06T09:17:18Z) - Training-Free Generalization on Heterogeneous Tabular Data via
Meta-Representation [67.30538142519067]
We propose Tabular data Pre-Training via Meta-representation (TabPTM)
A deep neural network is then trained to associate these meta-representations with dataset-specific classification confidences.
Experiments validate that TabPTM achieves promising performance in new datasets, even under few-shot scenarios.
arXiv Detail & Related papers (2023-10-31T18:03:54Z) - Generative Table Pre-training Empowers Models for Tabular Prediction [71.76829961276032]
We propose TapTap, the first attempt that leverages table pre-training to empower models for tabular prediction.
TapTap can generate high-quality synthetic tables to support various applications, including privacy protection, low resource regime, missing value imputation, and imbalanced classification.
It can be easily combined with various backbone models, including LightGBM, Multilayer Perceptron (MLP) and Transformer.
arXiv Detail & Related papers (2023-05-16T06:37:38Z) - Realistic Data Augmentation Framework for Enhancing Tabular Reasoning [15.339526664699845]
Existing approaches to constructing training data for Natural Language Inference tasks, such as for semi-structured table reasoning, are either via crowdsourcing or fully automatic methods.
This paper develops a realistic semi-automated framework for data augmentation for tabular inference.
arXiv Detail & Related papers (2022-10-23T17:32:19Z) - Understanding tables with intermediate pre-training [11.96734018295146]
We adapt TAPAS, a table-based BERT model, to recognize entailment.
We evaluate table pruning techniques as a pre-processing step to drastically improve the training and prediction efficiency.
arXiv Detail & Related papers (2020-10-01T17:43:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.