Transformer-based Language Models for Factoid Question Answering at
BioASQ9b
- URL: http://arxiv.org/abs/2109.07185v1
- Date: Wed, 15 Sep 2021 10:01:06 GMT
- Title: Transformer-based Language Models for Factoid Question Answering at
BioASQ9b
- Authors: Urvashi Khanna and Diego Moll\'a
- Abstract summary: We describe our experiments and participating systems in the BioASQ Task 9b Phase B challenge of biomedical question answering.
For factoid questions, our ALBERT-based systems ranked first in test batch 1 and fourth in test batch 2.
Our DistilBERT systems outperformed the ALBERT variants in test batches 4 and 5 despite having 81% fewer parameters than ALBERT.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we describe our experiments and participating systems in the
BioASQ Task 9b Phase B challenge of biomedical question answering. We have
focused on finding the ideal answers and investigated multi-task fine-tuning
and gradual unfreezing techniques on transformer-based language models. For
factoid questions, our ALBERT-based systems ranked first in test batch 1 and
fourth in test batch 2. Our DistilBERT systems outperformed the ALBERT variants
in test batches 4 and 5 despite having 81% fewer parameters than ALBERT.
However, we observed that gradual unfreezing had no significant impact on the
model's accuracy compared to standard fine-tuning.
Related papers
- oBERTa: Improving Sparse Transfer Learning via improved initialization,
distillation, and pruning regimes [82.99830498937729]
oBERTa is an easy-to-use set of language models for Natural Language Processing.
It allows NLP practitioners to obtain between 3.8 and 24.3 times faster models without expertise in model compression.
We explore the use of oBERTa on seven representative NLP tasks.
arXiv Detail & Related papers (2023-03-30T01:37:19Z) - Transformer-based approaches to Sentiment Detection [55.41644538483948]
We examined the performance of four different types of state-of-the-art transformer models for text classification.
The RoBERTa transformer model performs best on the test dataset with a score of 82.6% and is highly recommended for quality predictions.
arXiv Detail & Related papers (2023-03-13T17:12:03Z) - DELTA: degradation-free fully test-time adaptation [59.74287982885375]
We find that two unfavorable defects are concealed in the prevalent adaptation methodologies like test-time batch normalization (BN) and self-learning.
First, we reveal that the normalization statistics in test-time BN are completely affected by the currently received test samples, resulting in inaccurate estimates.
Second, we show that during test-time adaptation, the parameter update is biased towards some dominant classes.
arXiv Detail & Related papers (2023-01-30T15:54:00Z) - Understanding Prior Bias and Choice Paralysis in Transformer-based
Language Representation Models through Four Experimental Probes [8.591839265985412]
We present four confusion probes to test for problems such as prior bias and choice paralysis.
We show that the model exhibits significant prior bias and to a lesser, but still highly significant degree, choice paralysis, in addition to other problems.
Our results suggest that stronger testing protocols and additional benchmarks may be necessary before the language models are used in front-facing systems or decision making with real world consequences.
arXiv Detail & Related papers (2022-10-03T22:36:44Z) - Efficient Test-Time Model Adaptation without Forgetting [60.36499845014649]
Test-time adaptation seeks to tackle potential distribution shifts between training and testing data.
We propose an active sample selection criterion to identify reliable and non-redundant samples.
We also introduce a Fisher regularizer to constrain important model parameters from drastic changes.
arXiv Detail & Related papers (2022-04-06T06:39:40Z) - TernaryBERT: Distillation-aware Ultra-low Bit BERT [53.06741585060951]
We propose TernaryBERT, which ternarizes the weights in a fine-tuned BERT model.
Experiments on the GLUE benchmark and SQuAD show that our proposed TernaryBERT outperforms the other BERT quantization methods.
arXiv Detail & Related papers (2020-09-27T10:17:28Z) - Transferability of Natural Language Inference to Biomedical Question
Answering [17.38537039378825]
We focus on applying BioBERT to transfer the knowledge of natural language inference (NLI) to biomedical question answering (QA)
We observe that BioBERT trained on the NLI dataset obtains better performance on Yes/No (+5.59%), Factoid (+0.53%), List type (+13.58%) questions.
We present a sequential transfer learning method that significantly performed well in the 8th BioASQ Challenge (Phase B)
arXiv Detail & Related papers (2020-07-01T04:05:48Z) - DeeBERT: Dynamic Early Exiting for Accelerating BERT Inference [69.93692147242284]
Large-scale pre-trained language models such as BERT have brought significant improvements to NLP applications.
We propose a simple but effective method, DeeBERT, to accelerate BERT inference.
Experiments show that DeeBERT is able to save up to 40% inference time with minimal degradation in model quality.
arXiv Detail & Related papers (2020-04-27T17:58:05Z) - UNCC Biomedical Semantic Question Answering Systems. BioASQ: Task-7B,
Phase-B [1.976652238476722]
We present our approach for Task-7b, Phase B, Exact Answering Task.
These Question Answering (QA) tasks include Factoid, Yes/No, List Type Question answering.
Our system is based on a contextual word embedding model.
arXiv Detail & Related papers (2020-02-05T20:43:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.