TraceBack: Multi-Agent Decomposition for Fine-Grained Table Attribution
- URL: http://arxiv.org/abs/2602.13059v1
- Date: Fri, 13 Feb 2026 16:13:36 GMT
- Title: TraceBack: Multi-Agent Decomposition for Fine-Grained Table Attribution
- Authors: Tejas Anvekar, Junha Park, Rajat Jha, Devanshu Gupta, Poojah Ganesan, Puneeth Mathur, Vivek Gupta,
- Abstract summary: TraceBack is a framework for scalable, cell-level attribution in single-table QA.<n>We release CITEBench, a benchmark with phrase-to-cell annotations drawn from ToTTo, FetaQA, and AITQA.<n>We also propose FairScore, a reference-less metric that compares atomic facts derived from predicted cells and answers to estimate attribution precision and recall without human cell labels.
- Score: 11.133753556671392
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Question answering (QA) over structured tables requires not only accurate answers but also transparency about which cells support them. Existing table QA systems rarely provide fine-grained attribution, so even correct answers often lack verifiable grounding, limiting trust in high-stakes settings. We address this with TraceBack, a modular multi-agent framework for scalable, cell-level attribution in single-table QA. TraceBack prunes tables to relevant rows and columns, decomposes questions into semantically coherent sub-questions, and aligns each answer span with its supporting cells, capturing both explicit and implicit evidence used in intermediate reasoning steps. To enable systematic evaluation, we release CITEBench, a benchmark with phrase-to-cell annotations drawn from ToTTo, FetaQA, and AITQA. We further propose FairScore, a reference-less metric that compares atomic facts derived from predicted cells and answers to estimate attribution precision and recall without human cell labels. Experiments show that TraceBack substantially outperforms strong baselines across datasets and granularities, while FairScore closely tracks human judgments and preserves relative method rankings, supporting interpretable and scalable evaluation of table-based QA.
Related papers
- SQuARE: Structured Query & Adaptive Retrieval Engine For Tabular Formats [0.0]
SQuARE is a hybrid retrieval framework with sheet-level, complexity-aware routing.<n>It computes a continuous score based on header depth and merge density.<n>SQuARE consistently surpasses single-strategy baselines and ChatGPT-4o on both retrieval precision and end-to-end answer accuracy.
arXiv Detail & Related papers (2025-12-03T22:11:45Z) - Hierarchical Sequence Iteration for Heterogeneous Question Answering [27.22775290181187]
This paper introduces HSEQ, a unified framework that linearizes documents, tables, and knowledge graphs into a reversible hierarchical sequence.<n> Experiments on HotpotQA (text), HybridQA/TAT-QA (table+text), and MetaQA (KG) show consistent EM/F1 gains over strong single-pass, multi-hop, and agentic RAG baselines with high efficiency.
arXiv Detail & Related papers (2025-10-23T12:48:18Z) - RAG over Tables: Hierarchical Memory Index, Multi-Stage Retrieval, and Benchmarking [63.253294691180635]
In real-world scenarios, beyond pure text, a substantial amount of knowledge is stored in tables.<n>We first propose a table-corpora-aware RAG framework, named T-RAG, which consists of the hierarchical memory index, multi-stage retrieval, and graph-aware prompting.
arXiv Detail & Related papers (2025-04-02T04:24:41Z) - KET-QA: A Dataset for Knowledge Enhanced Table Question Answering [63.56707527868466]
We propose to use a knowledge base (KB) as the external knowledge source for TableQA.
Every question requires the integration of information from both the table and the sub-graph to be answered.
We design a retriever-reasoner structured pipeline model to extract pertinent information from the vast knowledge sub-graph.
arXiv Detail & Related papers (2024-05-13T18:26:32Z) - SQUARE: Automatic Question Answering Evaluation using Multiple Positive
and Negative References [73.67707138779245]
We propose a new evaluation metric: SQuArE (Sentence-level QUestion AnsweRing Evaluation)
We evaluate SQuArE on both sentence-level extractive (Answer Selection) and generative (GenQA) QA systems.
arXiv Detail & Related papers (2023-09-21T16:51:30Z) - Localize, Retrieve and Fuse: A Generalized Framework for Free-Form
Question Answering over Tables [46.039687237878105]
TableQA aims at generating answers to questions grounded on a provided table.
Table-to- Graph conversion, cell localizing, external knowledge retrieval, and the fusion of table and text are proposed.
Experiments showcase the superior capabilities of TAG-QA in generating sentences that are both faithful and coherent.
arXiv Detail & Related papers (2023-09-20T03:52:34Z) - TACR: A Table-alignment-based Cell-selection and Reasoning Model for
Hybrid Question-Answering [31.79113994947629]
We propose a Table-alignment-based Cell-selection and Reasoning model (TACR) for hybrid text and table QA.
In evidence retrieval, we design a table-question-alignment enhanced cell-selection method to retrieve fine-grained evidence.
In answer reasoning, we incorporate a QA module that treats the row containing selected cells as context.
arXiv Detail & Related papers (2023-05-24T03:42:44Z) - End-to-End Table Question Answering via Retrieval-Augmented Generation [19.89730342792824]
We introduce T-RAG, an end-to-end Table QA model, where a non-parametric dense vector index is fine-tuned jointly with BART, a parametric sequence-to-sequence model to generate answer tokens.
Given any natural language question, T-RAG utilizes a unified pipeline to automatically search through a table corpus to directly locate the correct answer from the table cells.
arXiv Detail & Related papers (2022-03-30T23:30:16Z) - Open Question Answering over Tables and Text [55.8412170633547]
In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Most open QA systems have considered only retrieving information from unstructured text.
We present a new large-scale dataset Open Table-and-Text Question Answering (OTT-QA) to evaluate performance on this task.
arXiv Detail & Related papers (2020-10-20T16:48:14Z) - Robust Question Answering Through Sub-part Alignment [53.94003466761305]
We model question answering as an alignment problem.
We train our model on SQuAD v1.1 and test it on several adversarial and out-of-domain datasets.
arXiv Detail & Related papers (2020-04-30T09:10:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.