Answering Ambiguous Questions through Generative Evidence Fusion and
Round-Trip Prediction
- URL: http://arxiv.org/abs/2011.13137v2
- Date: Sun, 30 May 2021 07:07:19 GMT
- Title: Answering Ambiguous Questions through Generative Evidence Fusion and
Round-Trip Prediction
- Authors: Yifan Gao, Henghui Zhu, Patrick Ng, Cicero Nogueira dos Santos, Zhiguo
Wang, Feng Nan, Dejiao Zhang, Ramesh Nallapati, Andrew O. Arnold, Bing Xiang
- Abstract summary: We present a model that aggregates and combines evidence from multiple passages to adaptively predict a single answer or a set of question-answer pairs for ambiguous questions.
Our model, named Refuel, achieves a new state-of-the-art performance on the AmbigQA dataset, and shows competitive performance on NQ-Open and TriviaQA.
- Score: 46.38201136570501
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In open-domain question answering, questions are highly likely to be
ambiguous because users may not know the scope of relevant topics when
formulating them. Therefore, a system needs to find possible interpretations of
the question, and predict one or multiple plausible answers. When multiple
plausible answers are found, the system should rewrite the question for each
answer to resolve the ambiguity. In this paper, we present a model that
aggregates and combines evidence from multiple passages to adaptively predict a
single answer or a set of question-answer pairs for ambiguous questions. In
addition, we propose a novel round-trip prediction approach to iteratively
generate additional interpretations that our model fails to find in the first
pass, and then verify and filter out the incorrect question-answer pairs to
arrive at the final disambiguated output. Our model, named Refuel, achieves a
new state-of-the-art performance on the AmbigQA dataset, and shows competitive
performance on NQ-Open and TriviaQA. The proposed round-trip prediction is a
model-agnostic general approach for answering ambiguous open-domain questions,
which improves our Refuel as well as several baseline models. We release source
code for our models and experiments at
https://github.com/amzn/refuel-open-domain-qa.
Related papers
- Answering Ambiguous Questions via Iterative Prompting [84.3426020642704]
In open-domain question answering, due to the ambiguity of questions, multiple plausible answers may exist.
One approach is to directly predict all valid answers, but this can struggle with balancing relevance and diversity.
We present AmbigPrompt to address the imperfections of existing approaches to answering ambiguous questions.
arXiv Detail & Related papers (2023-07-08T04:32:17Z) - Model Analysis & Evaluation for Ambiguous Question Answering [0.0]
Question Answering models are required to generate long-form answers that often combine conflicting pieces of information.
Recent advances in the field have shown strong capabilities in generating fluent responses, but certain research questions remain unanswered.
We aim to thoroughly investigate these aspects, and provide valuable insights into the limitations of the current approaches.
arXiv Detail & Related papers (2023-05-21T15:20:20Z) - Realistic Conversational Question Answering with Answer Selection based
on Calibrated Confidence and Uncertainty Measurement [54.55643652781891]
Conversational Question Answering (ConvQA) models aim at answering a question with its relevant paragraph and previous question-answer pairs that occurred during conversation multiple times.
We propose to filter out inaccurate answers in the conversation history based on their estimated confidences and uncertainties from the ConvQA model.
We validate our models, Answer Selection-based realistic Conversation Question Answering, on two standard ConvQA datasets.
arXiv Detail & Related papers (2023-02-10T09:42:07Z) - CREPE: Open-Domain Question Answering with False Presuppositions [92.20501870319765]
We introduce CREPE, a QA dataset containing a natural distribution of presupposition failures from online information-seeking forums.
We find that 25% of questions contain false presuppositions, and provide annotations for these presuppositions and their corrections.
We show that adaptations of existing open-domain QA models can find presuppositions moderately well, but struggle when predicting whether a presupposition is factually correct.
arXiv Detail & Related papers (2022-11-30T18:54:49Z) - A Wrong Answer or a Wrong Question? An Intricate Relationship between
Question Reformulation and Answer Selection in Conversational Question
Answering [15.355557454305776]
We show that question rewriting (QR) of the conversational context allows to shed more light on this phenomenon.
We present the results of this analysis on the TREC CAsT and QuAC (CANARD) datasets.
arXiv Detail & Related papers (2020-10-13T06:29:51Z) - Robust Question Answering Through Sub-part Alignment [53.94003466761305]
We model question answering as an alignment problem.
We train our model on SQuAD v1.1 and test it on several adversarial and out-of-domain datasets.
arXiv Detail & Related papers (2020-04-30T09:10:57Z) - AmbigQA: Answering Ambiguous Open-domain Questions [99.59747941602684]
We introduce AmbigQA, a new open-domain question answering task which involves finding every plausible answer.
To study this task, we construct AmbigNQ, a dataset covering 14,042 questions from NQ-open.
We find that over half of the questions in NQ-open are ambiguous, with diverse sources of ambiguity such as event and entity references.
arXiv Detail & Related papers (2020-04-22T15:42:13Z) - Rephrasing visual questions by specifying the entropy of the answer
distribution [0.0]
We propose a novel task, rephrasing the questions by controlling the ambiguity of the questions.
The ambiguity of a visual question is defined by the use of the entropy of the answer distribution predicted by a VQA model.
We demonstrate the advantage of our approach that can control the ambiguity of the rephrased questions, and an interesting observation that it is harder to increase than to reduce ambiguity.
arXiv Detail & Related papers (2020-04-10T09:32:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.