Sequence Tagging for Biomedical Extractive Question Answering
- URL: http://arxiv.org/abs/2104.07535v1
- Date: Thu, 15 Apr 2021 15:42:34 GMT
- Title: Sequence Tagging for Biomedical Extractive Question Answering
- Authors: Wonjin Yoon, Richard Jackson, Jaewoo Kang, Aron Lagerberg
- Abstract summary: We investigate the difference of the question distribution across the general and biomedical domains.
We discover biomedical questions are more likely to require list-type answers (multiple answers) than factoid-type answers (single answer)
Our approach can learn to decide the number of answers for a question from training data.
- Score: 12.464143741310137
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current studies in extractive question answering (EQA) have modeled
single-span extraction setting, where a single answer span is a label to
predict for a given question-passage pair. This setting is natural for general
domain EQA as the majority of the questions in the general domain can be
answered with a single span. Following general domain EQA models, current
biomedical EQA (BioEQA) models utilize single-span extraction setting with
post-processing steps. In this paper, we investigate the difference of the
question distribution across the general and biomedical domains and discover
biomedical questions are more likely to require list-type answers (multiple
answers) than factoid-type answers (single answer). In real-world use cases,
this emphasizes the need for Biomedical EQA models able to handle multiple
question types. Based on this preliminary study, we propose a multi-span
extraction setting, namely sequence tagging approach for BioEQA, which directly
tackles questions with a variable number of phrases as their answer. Our
approach can learn to decide the number of answers for a question from training
data. Our experimental result on the BioASQ 7b and 8b list-type questions
outperformed the best-performing existing models without requiring
post-processing steps.
Related papers
- Multi-LLM QA with Embodied Exploration [55.581423861790945]
We investigate the use of Multi-Embodied LLM Explorers (MELE) for question-answering in an unknown environment.
Multiple LLM-based agents independently explore and then answer queries about a household environment.
We analyze different aggregation methods to generate a single, final answer for each query.
arXiv Detail & Related papers (2024-06-16T12:46:40Z) - GSQA: An End-to-End Model for Generative Spoken Question Answering [54.418723701886115]
We introduce the first end-to-end Generative Spoken Question Answering (GSQA) model that empowers the system to engage in abstractive reasoning.
Our model surpasses the previous extractive model by 3% on extractive QA datasets.
Our GSQA model shows the potential to generalize to a broad spectrum of questions, thus further expanding the spoken question answering capabilities of abstractive QA.
arXiv Detail & Related papers (2023-12-15T13:33:18Z) - SEMQA: Semi-Extractive Multi-Source Question Answering [94.04430035121136]
We introduce a new QA task for answering multi-answer questions by summarizing multiple diverse sources in a semi-extractive fashion.
We create the first dataset of this kind, QuoteSum, with human-written semi-extractive answers to natural and generated questions.
arXiv Detail & Related papers (2023-11-08T18:46:32Z) - Adapting Pre-trained Generative Models for Extractive Question Answering [4.993041970406846]
We introduce a novel approach that uses the power of pre-trained generative models to address extractive QA tasks.
We demonstrate the superior performance of our proposed approach compared to existing state-of-the-art models.
arXiv Detail & Related papers (2023-11-06T09:01:02Z) - An Empirical Comparison of LM-based Question and Answer Generation
Methods [79.31199020420827]
Question and answer generation (QAG) consists of generating a set of question-answer pairs given a context.
In this paper, we establish baselines with three different QAG methodologies that leverage sequence-to-sequence language model (LM) fine-tuning.
Experiments show that an end-to-end QAG model, which is computationally light at both training and inference times, is generally robust and outperforms other more convoluted approaches.
arXiv Detail & Related papers (2023-05-26T14:59:53Z) - Activity report analysis with automatic single or multispan answer
extraction [0.21485350418225244]
We create a new smart home environment dataset comprised of questions paired with single-span or multi-span answers depending on the question and context queried.
Our experiments show that the proposed model outperforms state-of-the-art QA models on our dataset.
arXiv Detail & Related papers (2022-09-09T06:33:29Z) - BioTABQA: Instruction Learning for Biomedical Table Question Answering [19.66452178704578]
Table Question Answering (TQA) is an important but under-explored task.
None of TQA datasets exist in the biomedical domain where tables are frequently used to present information.
BioTABQA can not only be used to teach a model how to answer questions from tables but also evaluate how a model generalizes to unseen questions.
arXiv Detail & Related papers (2022-07-06T03:40:10Z) - Mixture of Experts for Biomedical Question Answering [34.92691831878302]
We propose a Mixture-of-Expert (MoE) based question answering method called MoEBQA.
MoEBQA decouples the computation for different types of questions by sparse routing.
We evaluate MoEBQA on three Biomedical Question Answering (BQA) datasets constructed based on real examinations.
arXiv Detail & Related papers (2022-04-15T14:11:40Z) - MixQG: Neural Question Generation with Mixed Answer Types [54.23205265351248]
We propose a neural question generator, MixQG, to bridge this gap.
We combine 9 question answering datasets with diverse answer types, including yes/no, multiple-choice, extractive, and abstractive answers.
Our model outperforms existing work in both seen and unseen domains.
arXiv Detail & Related papers (2021-10-15T16:03:40Z) - GooAQ: Open Question Answering with Diverse Answer Types [63.06454855313667]
We present GooAQ, a large-scale dataset with a variety of answer types.
This dataset contains over 5 million questions and 3 million answers collected from Google.
arXiv Detail & Related papers (2021-04-18T05:40:39Z) - Biomedical Question Answering: A Comprehensive Review [19.38459023509541]
Question Answering (QA) is a benchmark Natural Language Processing (NLP) task where models predict the answer for a given question using related documents, images, knowledge bases and question-answer pairs.
For specific domains like biomedicine, QA systems are still rarely used in real-life settings.
Biomedical QA (BQA), as an emerging QA task, enables innovative applications to effectively perceive, access and understand complex biomedical knowledge.
arXiv Detail & Related papers (2021-02-10T06:16:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.