Probabilistic Assumptions Matter: Improved Models for
Distantly-Supervised Document-Level Question Answering
- URL: http://arxiv.org/abs/2005.01898v1
- Date: Tue, 5 May 2020 01:08:36 GMT
- Title: Probabilistic Assumptions Matter: Improved Models for
Distantly-Supervised Document-Level Question Answering
- Authors: Hao Cheng, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
- Abstract summary: We address the problem of extractive question answering using document-level distant super-vision.
We show that these assumptions interact, and that different configurations provide complementary benefits.
Our approach outperforms previous state-of-the-art models by 4.3 points in F1 on TriviaQA-Wiki and 1.7 points in Rouge-L on NarrativeQA summaries.
- Score: 35.55031325165487
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address the problem of extractive question answering using document-level
distant super-vision, pairing questions and relevant documents with answer
strings. We compare previously used probability space and distant super-vision
assumptions (assumptions on the correspondence between the weak answer string
labels and possible answer mention spans). We show that these assumptions
interact, and that different configurations provide complementary benefits. We
demonstrate that a multi-objective model can efficiently combine the advantages
of multiple assumptions and out-perform the best individual formulation. Our
approach outperforms previous state-of-the-art models by 4.3 points in F1 on
TriviaQA-Wiki and 1.7 points in Rouge-L on NarrativeQA summaries.
Related papers
- Long-Span Question-Answering: Automatic Question Generation and QA-System Ranking via Side-by-Side Evaluation [65.16137964758612]
We explore the use of long-context capabilities in large language models to create synthetic reading comprehension data from entire books.
Our objective is to test the capabilities of LLMs to analyze, understand, and reason over problems that require a detailed comprehension of long spans of text.
arXiv Detail & Related papers (2024-05-31T20:15:10Z) - Improve Query Focused Abstractive Summarization by Incorporating Answer
Relevance [43.820971952979875]
We propose QFS-BART, a model that incorporates the explicit answer relevance of the source documents given the query via a question answering model.
Our model can take advantage of large pre-trained models which improve the summarization performance significantly.
Empirical results on the Debatepedia dataset show that the proposed model achieves the new state-of-the-art performance.
arXiv Detail & Related papers (2021-05-27T06:58:42Z) - Generative Context Pair Selection for Multi-hop Question Answering [60.74354009152721]
We propose a generative context selection model for multi-hop question answering.
Our proposed generative passage selection model has a better performance (4.9% higher than baseline) on adversarial held-out set.
arXiv Detail & Related papers (2021-04-18T07:00:48Z) - Learning with Instance Bundles for Reading Comprehension [61.823444215188296]
We introduce new supervision techniques that compare question-answer scores across multiple related instances.
Specifically, we normalize these scores across various neighborhoods of closely contrasting questions and/or answers.
We empirically demonstrate the effectiveness of training with instance bundles on two datasets.
arXiv Detail & Related papers (2021-04-18T06:17:54Z) - Mitigating False-Negative Contexts in Multi-document QuestionAnswering
with Retrieval Marginalization [29.797379277423143]
We develop a new parameterization of set-valued retrieval that properly handles unanswerable queries.
We show that marginalizing over this set during training allows a model to mitigate false negatives in annotated supporting evidences.
On IIRC, we show that joint modeling with marginalization on alternative contexts improves model performance by 5.5 F1 points and achieves a new state-of-the-art performance of 50.6 F1.
arXiv Detail & Related papers (2021-03-22T23:44:35Z) - Answering Ambiguous Questions through Generative Evidence Fusion and
Round-Trip Prediction [46.38201136570501]
We present a model that aggregates and combines evidence from multiple passages to adaptively predict a single answer or a set of question-answer pairs for ambiguous questions.
Our model, named Refuel, achieves a new state-of-the-art performance on the AmbigQA dataset, and shows competitive performance on NQ-Open and TriviaQA.
arXiv Detail & Related papers (2020-11-26T05:48:55Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z) - Robust Question Answering Through Sub-part Alignment [53.94003466761305]
We model question answering as an alignment problem.
We train our model on SQuAD v1.1 and test it on several adversarial and out-of-domain datasets.
arXiv Detail & Related papers (2020-04-30T09:10:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.