Transferable Student Performance Modeling for Intelligent Tutoring
Systems
- URL: http://arxiv.org/abs/2202.03980v1
- Date: Tue, 8 Feb 2022 16:36:27 GMT
- Title: Transferable Student Performance Modeling for Intelligent Tutoring
Systems
- Authors: Robin Schmucker, Tom M. Mitchell
- Abstract summary: We consider transfer learning techniques as a way to provide accurate performance predictions for new courses by leveraging log data from existing courses.
We evaluate the proposed techniques using student interaction sequence data from 5 different mathematics courses containing data from over 47,000 students in a real world large-scale ITS.
- Score: 24.118429574890055
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Millions of learners worldwide are now using intelligent tutoring systems
(ITSs). At their core, ITSs rely on machine learning algorithms to track each
user's changing performance level over time to provide personalized
instruction. Crucially, student performance models are trained using
interaction sequence data of previous learners to analyse data generated by
future learners. This induces a cold-start problem when a new course is
introduced for which no training data is available. Here, we consider transfer
learning techniques as a way to provide accurate performance predictions for
new courses by leveraging log data from existing courses. We study two
settings: (i) In the naive transfer setting, we propose course-agnostic
performance models that can be applied to any course. (ii) In the inductive
transfer setting, we tune pre-trained course-agnostic performance models to new
courses using small-scale target course data (e.g., collected during a pilot
study). We evaluate the proposed techniques using student interaction sequence
data from 5 different mathematics courses containing data from over 47,000
students in a real world large-scale ITS. The course-agnostic models that use
additional features provided by human domain experts (e.g, difficulty ratings
for questions in the new course) but no student interaction training data for
the new course, achieve prediction accuracy on par with standard BKT and PFA
models that use training data from thousands of students in the new course. In
the inductive setting our transfer learning approach yields more accurate
predictions than conventional performance models when only limited student
interaction training data (<100 students) is available to both.
Related papers
- Towards An Online Incremental Approach to Predict Students Performance [0.8287206589886879]
We propose a memory-based online incremental learning approach for updating an online classifier.
Our approach achieves a notable improvement in model accuracy, with an enhancement of nearly 10% compared to the current state-of-the-art.
arXiv Detail & Related papers (2024-05-03T17:13:26Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Robust Machine Learning by Transforming and Augmenting Imperfect
Training Data [6.928276018602774]
This thesis explores several data sensitivities of modern machine learning.
We first discuss how to prevent ML from codifying prior human discrimination measured in the training data.
We then discuss the problem of learning from data containing spurious features, which provide predictive fidelity during training but are unreliable upon deployment.
arXiv Detail & Related papers (2023-12-19T20:49:28Z) - Efficient Grammatical Error Correction Via Multi-Task Training and
Optimized Training Schedule [55.08778142798106]
We propose auxiliary tasks that exploit the alignment between the original and corrected sentences.
We formulate each task as a sequence-to-sequence problem and perform multi-task training.
We find that the order of datasets used for training and even individual instances within a dataset may have important effects on the final performance.
arXiv Detail & Related papers (2023-11-20T14:50:12Z) - PILOT: A Pre-Trained Model-Based Continual Learning Toolbox [71.63186089279218]
This paper introduces a pre-trained model-based continual learning toolbox known as PILOT.
On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt.
On the other hand, PILOT fits typical class-incremental learning algorithms within the context of pre-trained models to evaluate their effectiveness.
arXiv Detail & Related papers (2023-09-13T17:55:11Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Process-BERT: A Framework for Representation Learning on Educational
Process Data [68.8204255655161]
We propose a framework for learning representations of educational process data.
Our framework consists of a pre-training step that uses BERT-type objectives to learn representations from sequential process data.
We apply our framework to the 2019 nation's report card data mining competition dataset.
arXiv Detail & Related papers (2022-04-28T16:07:28Z) - Assessing the Knowledge State of Online Students -- New Data, New
Approaches, Improved Accuracy [28.719009375724028]
Student performance (SP) modeling is a critical step for building adaptive online teaching systems.
This study is the first to use four very large datasets made available recently from four distinct intelligent tutoring systems.
arXiv Detail & Related papers (2021-09-04T00:08:59Z) - Federated Few-Shot Learning with Adversarial Learning [30.905239262227]
We propose a few-shot learning framework to learn a few-shot classification model that can classify unseen data classes with only a few labeled samples.
We show our approaches outperform baselines by more than 10% in learning vision tasks and 5% in language tasks.
arXiv Detail & Related papers (2021-04-01T09:44:57Z) - A Survey on Self-supervised Pre-training for Sequential Transfer
Learning in Neural Networks [1.1802674324027231]
Self-supervised pre-training for transfer learning is becoming an increasingly popular technique to improve state-of-the-art results using unlabeled data.
We provide an overview of the taxonomy for self-supervised learning and transfer learning, and highlight some prominent methods for designing pre-training tasks across different domains.
arXiv Detail & Related papers (2020-07-01T22:55:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.