Answering Ambiguous Questions with a Database of Questions, Answers, and
Revisions
- URL: http://arxiv.org/abs/2308.08661v1
- Date: Wed, 16 Aug 2023 20:23:16 GMT
- Title: Answering Ambiguous Questions with a Database of Questions, Answers, and
Revisions
- Authors: Haitian Sun, William W. Cohen, Ruslan Salakhutdinov
- Abstract summary: We present a new state-of-the-art for answering ambiguous questions that exploits a database of unambiguous questions generated from Wikipedia.
Our method improves performance by 15% on recall measures and 10% on measures which evaluate disambiguating questions from predicted outputs.
- Score: 95.92276099234344
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Many open-domain questions are under-specified and thus have multiple
possible answers, each of which is correct under a different interpretation of
the question. Answering such ambiguous questions is challenging, as it requires
retrieving and then reasoning about diverse information from multiple passages.
We present a new state-of-the-art for answering ambiguous questions that
exploits a database of unambiguous questions generated from Wikipedia. On the
challenging ASQA benchmark, which requires generating long-form answers that
summarize the multiple answers to an ambiguous question, our method improves
performance by 15% (relative improvement) on recall measures and 10% on
measures which evaluate disambiguating questions from predicted outputs.
Retrieving from the database of generated questions also gives large
improvements in diverse passage retrieval (by matching user questions q to
passages p indirectly, via questions q' generated from p).
Related papers
- Detecting Temporal Ambiguity in Questions [16.434748534272014]
Temporally ambiguous questions are one of the most common types of such questions.
Our annotations focus on capturing temporal ambiguity to study the task of detecting temporally ambiguous questions.
We propose a novel approach by using diverse search strategies based on disambiguated versions of the questions.
arXiv Detail & Related papers (2024-09-25T15:59:58Z) - Which questions should I answer? Salience Prediction of Inquisitive Questions [118.097974193544]
We show that highly salient questions are empirically more likely to be answered in the same article.
We further validate our findings by showing that answering salient questions is an indicator of summarization quality in news.
arXiv Detail & Related papers (2024-04-16T21:33:05Z) - Answering Unanswered Questions through Semantic Reformulations in Spoken
QA [20.216161323866867]
Spoken Question Answering (QA) is a key feature of voice assistants, usually backed by multiple QA systems.
We analyze failed QA requests to identify core challenges: lexical gaps, proposition types, complex syntactic structure, and high specificity.
We propose a Semantic Question Reformulation (SURF) model offering three linguistically-grounded operations (repair, syntactic reshaping, generalization) to rewrite questions to facilitate answering.
arXiv Detail & Related papers (2023-05-27T07:19:27Z) - Asking Clarification Questions to Handle Ambiguity in Open-Domain QA [25.80369529145732]
We propose to ask a clarification question, where the user's response will help identify the interpretation that best aligns with the user's intention.
We first present CAMBIGNQ, a dataset consisting of 5,654 ambiguous questions.
We then define a pipeline of tasks and design appropriate evaluation metrics.
arXiv Detail & Related papers (2023-05-23T08:20:01Z) - CREPE: Open-Domain Question Answering with False Presuppositions [92.20501870319765]
We introduce CREPE, a QA dataset containing a natural distribution of presupposition failures from online information-seeking forums.
We find that 25% of questions contain false presuppositions, and provide annotations for these presuppositions and their corrections.
We show that adaptations of existing open-domain QA models can find presuppositions moderately well, but struggle when predicting whether a presupposition is factually correct.
arXiv Detail & Related papers (2022-11-30T18:54:49Z) - QAMPARI: An Open-domain Question Answering Benchmark for Questions with
Many Answers from Multiple Paragraphs [42.843866049949206]
We introduce QAMPARI, an ODQA benchmark where question answers are lists of entities, spread across many paragraphs.
We create QAMPARI by (a) generating questions with multiple answers from Wikipedia's knowledge graph and tables, (b) automatically pairing answers with supporting evidence in Wikipedia paragraphs, and (c) manually paraphrasing questions and validating each answer.
We train ODQA models from the retrieve-and-read family and find that QAMPARI is challenging in terms of both passage retrieval and answer generation, reaching an F1 score of 32.8 at best.
arXiv Detail & Related papers (2022-05-25T11:21:30Z) - Discourse Comprehension: A Question Answering Framework to Represent
Sentence Connections [35.005593397252746]
A key challenge in building and evaluating models for discourse comprehension is the lack of annotated data.
This paper presents a novel paradigm that enables scalable data collection targeting the comprehension of news documents.
The resulting corpus, DCQA, consists of 22,430 question-answer pairs across 607 English documents.
arXiv Detail & Related papers (2021-11-01T04:50:26Z) - A Dataset of Information-Seeking Questions and Answers Anchored in
Research Papers [66.11048565324468]
We present a dataset of 5,049 questions over 1,585 Natural Language Processing papers.
Each question is written by an NLP practitioner who read only the title and abstract of the corresponding paper, and the question seeks information present in the full text.
We find that existing models that do well on other QA tasks do not perform well on answering these questions, underperforming humans by at least 27 F1 points when answering them from entire papers.
arXiv Detail & Related papers (2021-05-07T00:12:34Z) - GooAQ: Open Question Answering with Diverse Answer Types [63.06454855313667]
We present GooAQ, a large-scale dataset with a variety of answer types.
This dataset contains over 5 million questions and 3 million answers collected from Google.
arXiv Detail & Related papers (2021-04-18T05:40:39Z) - Unsupervised Question Decomposition for Question Answering [102.56966847404287]
We propose an algorithm for One-to-N Unsupervised Sequence Sequence (ONUS) that learns to map one hard, multi-hop question to many simpler, single-hop sub-questions.
We show large QA improvements on HotpotQA over a strong baseline on the original, out-of-domain, and multi-hop dev sets.
arXiv Detail & Related papers (2020-02-22T19:40:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.