The University of Texas at Dallas HLTRI's Participation in EPIC-QA:
Searching for Entailed Questions Revealing Novel Answer Nuggets
- URL: http://arxiv.org/abs/2112.13946v1
- Date: Tue, 28 Dec 2021 00:14:46 GMT
- Title: The University of Texas at Dallas HLTRI's Participation in EPIC-QA:
Searching for Entailed Questions Revealing Novel Answer Nuggets
- Authors: Maxwell Weinzierl, Sanda M. Harabagiu
- Abstract summary: This paper describes our participation in both tasks of EPIC-QA, targeting: (1) Expert QA and (2) Consumer QA.
Our methods used a multi-phase neural Information Retrieval (IR) system based on combining BM25, BERT, and T5 as well as the idea of considering entailment relations between the original question and questions automatically generated from answer candidate sentences.
Our system, called SEaRching for Entailed QUestions revealing NOVel nuggets of Answers (SER4EQUNOVA), produced promising results in both EPIC-QA tasks, excelling in the Expert QA task.
- Score: 1.0957528713294875
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Epidemic Question Answering (EPIC-QA) track at the Text Analysis
Conference (TAC) is an evaluation of methodologies for answering ad-hoc
questions about the COVID-19 disease. This paper describes our participation in
both tasks of EPIC-QA, targeting: (1) Expert QA and (2) Consumer QA. Our
methods used a multi-phase neural Information Retrieval (IR) system based on
combining BM25, BERT, and T5 as well as the idea of considering entailment
relations between the original question and questions automatically generated
from answer candidate sentences. Moreover, because entailment relations were
also considered between all generated questions, we were able to re-rank the
answer sentences based on the number of novel answer nuggets they contained, as
indicated by the processing of a question entailment graph. Our system, called
SEaRching for Entailed QUestions revealing NOVel nuggets of Answers
(SER4EQUNOVA), produced promising results in both EPIC-QA tasks, excelling in
the Expert QA task.
Related papers
- On the Robustness of Question Rewriting Systems to Questions of Varying
Hardness [43.63930447922717]
We are interested in the robustness of a QR system to questions varying in rewriting hardness or difficulty.
We first propose a method to automatically classify questions into subsets of varying hardness, by measuring the discrepancy between a question and its rewrite.
To enhance the robustness of QR systems to questions of varying hardness, we propose a novel learning framework for QR that first trains a QR model independently on each subset of questions of a certain level of hardness, then combines these QR models as one joint model for inference.
arXiv Detail & Related papers (2023-11-12T11:09:30Z) - SQUARE: Automatic Question Answering Evaluation using Multiple Positive
and Negative References [73.67707138779245]
We propose a new evaluation metric: SQuArE (Sentence-level QUestion AnsweRing Evaluation)
We evaluate SQuArE on both sentence-level extractive (Answer Selection) and generative (GenQA) QA systems.
arXiv Detail & Related papers (2023-09-21T16:51:30Z) - Answering Ambiguous Questions with a Database of Questions, Answers, and
Revisions [95.92276099234344]
We present a new state-of-the-art for answering ambiguous questions that exploits a database of unambiguous questions generated from Wikipedia.
Our method improves performance by 15% on recall measures and 10% on measures which evaluate disambiguating questions from predicted outputs.
arXiv Detail & Related papers (2023-08-16T20:23:16Z) - Answering Unanswered Questions through Semantic Reformulations in Spoken
QA [20.216161323866867]
Spoken Question Answering (QA) is a key feature of voice assistants, usually backed by multiple QA systems.
We analyze failed QA requests to identify core challenges: lexical gaps, proposition types, complex syntactic structure, and high specificity.
We propose a Semantic Question Reformulation (SURF) model offering three linguistically-grounded operations (repair, syntactic reshaping, generalization) to rewrite questions to facilitate answering.
arXiv Detail & Related papers (2023-05-27T07:19:27Z) - Consecutive Question Generation via Dynamic Multitask Learning [17.264399861776187]
We propose the task of consecutive question generation (CQG), which generates a set of logically related question-answer pairs to understand a whole passage.
We first examine the four key elements of CQG, and propose a novel dynamic multitask framework with one main task generating a question-answer pair, and four auxiliary tasks generating other elements.
We prove that our strategy can improve question generation significantly and benefit multiple related NLP tasks.
arXiv Detail & Related papers (2022-11-16T11:50:36Z) - PACIFIC: Towards Proactive Conversational Question Answering over
Tabular and Textual Data in Finance [96.06505049126345]
We present a new dataset, named PACIFIC. Compared with existing CQA datasets, PACIFIC exhibits three key features: (i) proactivity, (ii) numerical reasoning, and (iii) hybrid context of tables and text.
A new task is defined accordingly to study Proactive Conversational Question Answering (PCQA), which combines clarification question generation and CQA.
UniPCQA performs multi-task learning over all sub-tasks in PCQA and incorporates a simple ensemble strategy to alleviate the error propagation issue in the multi-task learning by cross-validating top-$k$ sampled Seq2Seq
arXiv Detail & Related papers (2022-10-17T08:06:56Z) - Generating Self-Contained and Summary-Centric Question Answer Pairs via
Differentiable Reward Imitation Learning [7.2745835227138045]
We propose a model for generating question-answer pairs (QA pairs) with self-contained, summary-centric questions and length-constrained, article-summarizing answers.
This dataset is used to learn a QA pair generation model producing summaries as answers that balance brevity with sufficiency jointly with their corresponding questions.
arXiv Detail & Related papers (2021-09-10T06:34:55Z) - A Dataset of Information-Seeking Questions and Answers Anchored in
Research Papers [66.11048565324468]
We present a dataset of 5,049 questions over 1,585 Natural Language Processing papers.
Each question is written by an NLP practitioner who read only the title and abstract of the corresponding paper, and the question seeks information present in the full text.
We find that existing models that do well on other QA tasks do not perform well on answering these questions, underperforming humans by at least 27 F1 points when answering them from entire papers.
arXiv Detail & Related papers (2021-05-07T00:12:34Z) - GooAQ: Open Question Answering with Diverse Answer Types [63.06454855313667]
We present GooAQ, a large-scale dataset with a variety of answer types.
This dataset contains over 5 million questions and 3 million answers collected from Google.
arXiv Detail & Related papers (2021-04-18T05:40:39Z) - A Wrong Answer or a Wrong Question? An Intricate Relationship between
Question Reformulation and Answer Selection in Conversational Question
Answering [15.355557454305776]
We show that question rewriting (QR) of the conversational context allows to shed more light on this phenomenon.
We present the results of this analysis on the TREC CAsT and QuAC (CANARD) datasets.
arXiv Detail & Related papers (2020-10-13T06:29:51Z) - Unsupervised Question Decomposition for Question Answering [102.56966847404287]
We propose an algorithm for One-to-N Unsupervised Sequence Sequence (ONUS) that learns to map one hard, multi-hop question to many simpler, single-hop sub-questions.
We show large QA improvements on HotpotQA over a strong baseline on the original, out-of-domain, and multi-hop dev sets.
arXiv Detail & Related papers (2020-02-22T19:40:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.