Question Generation for Adaptive Education
- URL: http://arxiv.org/abs/2106.04262v1
- Date: Tue, 8 Jun 2021 11:46:59 GMT
- Title: Question Generation for Adaptive Education
- Authors: Megha Srivastava and Noah Goodman
- Abstract summary: We show how to fine-tune pre-trained language models for deep knowledge tracing (LM-KT)
This model accurately predicts the probability of a student answering a question correctly, and generalizes to questions not seen in training.
We then use LM-KT to specify the objective and data for training a model to generate questions conditioned on the student and target difficulty.
- Score: 7.23389716633927
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Intelligent and adaptive online education systems aim to make high-quality
education available for a diverse range of students. However, existing systems
usually depend on a pool of hand-made questions, limiting how fine-grained and
open-ended they can be in adapting to individual students. We explore targeted
question generation as a controllable sequence generation task. We first show
how to fine-tune pre-trained language models for deep knowledge tracing
(LM-KT). This model accurately predicts the probability of a student answering
a question correctly, and generalizes to questions not seen in training. We
then use LM-KT to specify the objective and data for training a model to
generate questions conditioned on the student and target difficulty. Our
results show we succeed at generating novel, well-calibrated language
translation questions for second language learners from a real online education
platform.
Related papers
- Toward In-Context Teaching: Adapting Examples to Students' Misconceptions [54.82965010592045]
We introduce a suite of models and evaluation methods we call AdapT.
AToM is a new probabilistic model for adaptive teaching that jointly infers students' past beliefs and optimize for the correctness of future beliefs.
Our results highlight both the difficulty of the adaptive teaching task and the potential of learned adaptive models for solving it.
arXiv Detail & Related papers (2024-05-07T17:05:27Z) - Integrating A.I. in Higher Education: Protocol for a Pilot Study with 'SAMCares: An Adaptive Learning Hub' [0.6990493129893112]
This research aims to introduce an innovative study buddy we will be calling the 'SAMCares'
The system leverages a Large Language Model (LLM) and Retriever-Augmented Generation (RAG) to offer real-time, context-aware, and adaptive educational support.
arXiv Detail & Related papers (2024-05-01T05:39:07Z) - Adapting Large Language Models for Education: Foundational Capabilities, Potentials, and Challenges [60.62904929065257]
Large language models (LLMs) offer possibility for resolving this issue by comprehending individual requests.
This paper reviews the recently emerged LLM research related to educational capabilities, including mathematics, writing, programming, reasoning, and knowledge-based question answering.
arXiv Detail & Related papers (2023-12-27T14:37:32Z) - Preference-grounded Token-level Guidance for Language Model Fine-tuning [105.88789610320426]
Aligning language models with preferences is an important problem in natural language generation.
For LM training, based on the amount of supervised data, we present two *minimalist* learning objectives that utilize the learned guidance.
In experiments, our method performs competitively on two distinct representative LM tasks.
arXiv Detail & Related papers (2023-06-01T07:00:07Z) - Scalable Educational Question Generation with Pre-trained Language
Models [17.701780209575777]
We develop textitEduQG, a novel educational question generation model built by adapting a large language model.
Our experiments demonstrate that textitEduQG can produce superior educational questions by further pre-training and fine-tuning a pre-trained language model on the scientific text and science question data.
arXiv Detail & Related papers (2023-05-13T09:08:27Z) - Pre-Training With Scientific Text Improves Educational Question
Generation [17.701780209575777]
We develop EduQG, a novel educational question generation model built by adapting a large language model.
Our initial experiments demonstrate that EduQG can produce superior educational questions by pre-training on scientific text.
arXiv Detail & Related papers (2022-12-07T17:17:58Z) - Automatic Short Math Answer Grading via In-context Meta-learning [2.0263791972068628]
We study the problem of automatic short answer grading for students' responses to math questions.
We use MathBERT, a variant of the popular language model BERT adapted to mathematical content, as our base model.
Second, we use an in-context learning approach that provides scoring examples as input to the language model.
arXiv Detail & Related papers (2022-05-30T16:26:02Z) - Educational Question Generation of Children Storybooks via Question Type Distribution Learning and Event-Centric Summarization [67.1483219601714]
We propose a novel question generation method that first learns the question type distribution of an input story paragraph.
We finetune a pre-trained transformer-based sequence-to-sequence model using silver samples composed by educational question-answer pairs.
Our work indicates the necessity of decomposing question type distribution learning and event-centric summary generation for educational question generation.
arXiv Detail & Related papers (2022-03-27T02:21:19Z) - Exploring Bayesian Deep Learning for Urgent Instructor Intervention Need
in MOOC Forums [58.221459787471254]
Massive Open Online Courses (MOOCs) have become a popular choice for e-learning thanks to their great flexibility.
Due to large numbers of learners and their diverse backgrounds, it is taxing to offer real-time support.
With the large volume of posts and high workloads for MOOC instructors, it is unlikely that the instructors can identify all learners requiring intervention.
This paper explores for the first time Bayesian deep learning on learner-based text posts with two methods: Monte Carlo Dropout and Variational Inference.
arXiv Detail & Related papers (2021-04-26T15:12:13Z) - Inquisitive Question Generation for High Level Text Comprehension [60.21497846332531]
We introduce INQUISITIVE, a dataset of 19K questions that are elicited while a person is reading through a document.
We show that readers engage in a series of pragmatic strategies to seek information.
We evaluate question generation models based on GPT-2 and show that our model is able to generate reasonable questions.
arXiv Detail & Related papers (2020-10-04T19:03:39Z) - Educational Question Mining At Scale: Prediction, Analysis and
Personalization [35.42197158180065]
We propose a framework for mining insights from educational questions at scale.
We utilize the state-of-the-art Bayesian deep learning method, in particular partial variational auto-encoders (p-VAE)
We apply our proposed framework to a real-world dataset with tens of thousands of questions and tens of millions of answers from an online education platform.
arXiv Detail & Related papers (2020-03-12T19:07:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.