AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for
Skill Assessments
- URL: http://arxiv.org/abs/2012.01186v2
- Date: Fri, 18 Dec 2020 23:46:56 GMT
- Title: AGenT Zero: Zero-shot Automatic Multiple-Choice Question Generation for
Skill Assessments
- Authors: Eric Li, Jingyi Su, Hao Sheng, Lawrence Wai
- Abstract summary: Multiple-choice questions (MCQs) offer the most promising avenue for skill evaluation in the era of virtual education and job recruiting.
Recent advances in natural language processing have given rise to many complex question generation methods.
AGenT Zero successfully outperforms other pre-trained methods in fluency and semantic similarity.
- Score: 11.355397923795488
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiple-choice questions (MCQs) offer the most promising avenue for skill
evaluation in the era of virtual education and job recruiting, where
traditional performance-based alternatives such as projects and essays have
become less viable, and grading resources are constrained. The automated
generation of MCQs would allow assessment creation at scale. Recent advances in
natural language processing have given rise to many complex question generation
methods. However, the few methods that produce deployable results in specific
domains require a large amount of domain-specific training data that can be
very costly to acquire. Our work provides an initial foray into MCQ generation
under high data-acquisition cost scenarios by strategically emphasizing
paraphrasing the question context (compared to the task). In addition to
maintaining semantic similarity between the question-answer pairs, our
pipeline, which we call AGenT Zero, consists of only pre-trained models and
requires no fine-tuning, minimizing data acquisition costs for question
generation. AGenT Zero successfully outperforms other pre-trained methods in
fluency and semantic similarity. Additionally, with some small changes, our
assessment pipeline can be generalized to a broader question and answer space,
including short answer or fill in the blank questions.
Related papers
- Adapting Pre-trained Generative Models for Extractive Question Answering [4.993041970406846]
We introduce a novel approach that uses the power of pre-trained generative models to address extractive QA tasks.
We demonstrate the superior performance of our proposed approach compared to existing state-of-the-art models.
arXiv Detail & Related papers (2023-11-06T09:01:02Z) - A Lightweight Method to Generate Unanswerable Questions in English [18.323248259867356]
We examine a simpler data augmentation method for unanswerable question generation in English.
We perform antonym and entity swaps on answerable questions.
Compared to the prior state-of-the-art, data generated with our training-free and lightweight strategy results in better models.
arXiv Detail & Related papers (2023-10-30T10:14:52Z) - QASnowball: An Iterative Bootstrapping Framework for High-Quality
Question-Answering Data Generation [67.27999343730224]
We introduce an iterative bootstrapping framework for QA data augmentation (named QASnowball)
QASnowball can iteratively generate large-scale high-quality QA data based on a seed set of supervised examples.
We conduct experiments in the high-resource English scenario and the medium-resource Chinese scenario, and the experimental results show that the data generated by QASnowball can facilitate QA models.
arXiv Detail & Related papers (2023-09-19T05:20:36Z) - An Empirical Comparison of LM-based Question and Answer Generation
Methods [79.31199020420827]
Question and answer generation (QAG) consists of generating a set of question-answer pairs given a context.
In this paper, we establish baselines with three different QAG methodologies that leverage sequence-to-sequence language model (LM) fine-tuning.
Experiments show that an end-to-end QAG model, which is computationally light at both training and inference times, is generally robust and outperforms other more convoluted approaches.
arXiv Detail & Related papers (2023-05-26T14:59:53Z) - PAXQA: Generating Cross-lingual Question Answering Examples at Training
Scale [53.92008514395125]
PAXQA (Projecting annotations for cross-lingual (x) QA) decomposes cross-lingual QA into two stages.
We propose a novel use of lexically-constrained machine translation, in which constrained entities are extracted from the parallel bitexts.
We show that models fine-tuned on these datasets outperform prior synthetic data generation models over several extractive QA datasets.
arXiv Detail & Related papers (2023-04-24T15:46:26Z) - Momentum Contrastive Pre-training for Question Answering [54.57078061878619]
MCROSS introduces a momentum contrastive learning framework to align the answer probability between cloze-like and natural query-passage sample pairs.
Our method achieves noticeable improvement compared with all baselines in both supervised and zero-shot scenarios.
arXiv Detail & Related papers (2022-12-12T08:28:22Z) - Successive Prompting for Decomposing Complex Questions [50.00659445976735]
Recent works leverage the capabilities of large language models (LMs) to perform complex question answering in a few-shot setting.
We introduce Successive Prompting'', where we iteratively break down a complex task into a simple task, solve it, and then repeat the process until we get the final solution.
Our best model (with successive prompting) achieves an improvement of 5% absolute F1 on a few-shot version of the DROP dataset.
arXiv Detail & Related papers (2022-12-08T06:03:38Z) - Automatic Short Math Answer Grading via In-context Meta-learning [2.0263791972068628]
We study the problem of automatic short answer grading for students' responses to math questions.
We use MathBERT, a variant of the popular language model BERT adapted to mathematical content, as our base model.
Second, we use an in-context learning approach that provides scoring examples as input to the language model.
arXiv Detail & Related papers (2022-05-30T16:26:02Z) - Contrastive Domain Adaptation for Question Answering using Limited Text
Corpora [20.116147632481983]
We propose a novel framework for domain adaptation called contrastive domain adaptation for QA.
Specifically, CAQA combines techniques from question generation and domain-invariant learning to answer out-of-domain questions in settings with limited text corpora.
arXiv Detail & Related papers (2021-08-31T14:05:55Z) - ManyModalQA: Modality Disambiguation and QA over Diverse Inputs [73.93607719921945]
We present a new multimodal question answering challenge, ManyModalQA, in which an agent must answer a question by considering three distinct modalities.
We collect our data by scraping Wikipedia and then utilize crowdsourcing to collect question-answer pairs.
arXiv Detail & Related papers (2020-01-22T14:39:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.