simpleKT: A Simple But Tough-to-Beat Baseline for Knowledge Tracing
- URL: http://arxiv.org/abs/2302.06881v1
- Date: Tue, 14 Feb 2023 08:09:09 GMT
- Title: simpleKT: A Simple But Tough-to-Beat Baseline for Knowledge Tracing
- Authors: Zitao Liu, Qiongqiong Liu, Jiahao Chen, Shuyan Huang, Weiqi Luo
- Abstract summary: We provide a strong but simple baseline method to deal with the KT task named textscsimpleKT.
Inspired by the Rasch model in psychometrics, we explicitly model question-specific variations to capture the individual differences among questions.
We use the ordinary dot-product attention function to extract the time-aware information embedded in the student learning interactions.
- Score: 22.055683237994696
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Knowledge tracing (KT) is the problem of predicting students' future
performance based on their historical interactions with intelligent tutoring
systems. Recently, many works present lots of special methods for applying deep
neural networks to KT from different perspectives like model architecture,
adversarial augmentation and etc., which make the overall algorithm and system
become more and more complex. Furthermore, due to the lack of standardized
evaluation protocol \citep{liu2022pykt}, there is no widely agreed KT baselines
and published experimental comparisons become inconsistent and
self-contradictory, i.e., the reported AUC scores of DKT on ASSISTments2009
range from 0.721 to 0.821 \citep{minn2018deep,yeung2018addressing}. Therefore,
in this paper, we provide a strong but simple baseline method to deal with the
KT task named \textsc{simpleKT}. Inspired by the Rasch model in psychometrics,
we explicitly model question-specific variations to capture the individual
differences among questions covering the same set of knowledge components that
are a generalization of terms of concepts or skills needed for learners to
accomplish steps in a task or a problem. Furthermore, instead of using
sophisticated representations to capture student forgetting behaviors, we use
the ordinary dot-product attention function to extract the time-aware
information embedded in the student learning interactions. Extensive
experiments show that such a simple baseline is able to always rank top 3 in
terms of AUC scores and achieve 57 wins, 3 ties and 16 loss against 12 DLKT
baseline methods on 7 public datasets of different domains. We believe this
work serves as a strong baseline for future KT research. Code is available at
\url{https://github.com/pykt-team/pykt-toolkit}\footnote{We merged our model to
the \textsc{pyKT} benchmark at \url{https://pykt.org/}.}.
Related papers
- Automated Knowledge Concept Annotation and Question Representation Learning for Knowledge Tracing [59.480951050911436]
We present KCQRL, a framework for automated knowledge concept annotation and question representation learning.
We demonstrate the effectiveness of KCQRL across 15 KT algorithms on two large real-world Math learning datasets.
arXiv Detail & Related papers (2024-10-02T16:37:19Z) - Towards Robust Knowledge Tracing Models via k-Sparse Attention [33.02197868261949]
textscsparseKT is a simple yet effective framework to improve the robustness and generalization of the attention based DLKT approaches.
We show that our textscsparseKT is able to help attentional KT models get rid of irrelevant student interactions.
arXiv Detail & Related papers (2024-07-24T08:49:18Z) - M-Tuning: Prompt Tuning with Mitigated Label Bias in Open-Set Scenarios [103.6153593636399]
We propose a vision-language prompt tuning method with mitigated label bias (M-Tuning)
It introduces open words from the WordNet to extend the range of words forming the prompt texts from only closed-set label words to more, and thus prompts are tuned in a simulated open-set scenario.
Our method achieves the best performance on datasets with various scales, and extensive ablation studies also validate its effectiveness.
arXiv Detail & Related papers (2023-03-09T09:05:47Z) - Enhancing Deep Knowledge Tracing with Auxiliary Tasks [24.780533765606922]
We propose emphAT-DKT to improve the prediction performance of the original deep knowledge tracing model.
We conduct comprehensive experiments on three real-world educational datasets and compare the proposed approach to both deep sequential KT models and non-sequential models.
arXiv Detail & Related papers (2023-02-14T08:21:37Z) - Improving Interpretability of Deep Sequential Knowledge Tracing Models
with Question-centric Cognitive Representations [22.055683237994696]
We present QIKT, a question-centric interpretable KT model to address the above challenges.
The proposed QIKT approach explicitly models students' knowledge state variations at a fine-grained level.
It outperforms a wide range of deep learning based KT models in terms of prediction accuracy with better model interpretability.
arXiv Detail & Related papers (2023-02-14T08:14:30Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - pyKT: A Python Library to Benchmark Deep Learning based Knowledge
Tracing Models [46.05383477261115]
Knowledge tracing (KT) is the task of using students' historical learning interaction data to model their knowledge mastery over time.
DLKT approaches are still left somewhat unknown and proper measurement and analysis of these approaches remain a challenge.
We introduce a comprehensive python based benchmark platform, textscpyKT, to guarantee valid comparisons across DLKT methods.
arXiv Detail & Related papers (2022-06-23T02:42:47Z) - Learning with Neighbor Consistency for Noisy Labels [69.83857578836769]
We present a method for learning from noisy labels that leverages similarities between training examples in feature space.
We evaluate our method on datasets evaluating both synthetic (CIFAR-10, CIFAR-100) and realistic (mini-WebVision, Clothing1M, mini-ImageNet-Red) noise.
arXiv Detail & Related papers (2022-02-04T15:46:27Z) - Learning with Instance Bundles for Reading Comprehension [61.823444215188296]
We introduce new supervision techniques that compare question-answer scores across multiple related instances.
Specifically, we normalize these scores across various neighborhoods of closely contrasting questions and/or answers.
We empirically demonstrate the effectiveness of training with instance bundles on two datasets.
arXiv Detail & Related papers (2021-04-18T06:17:54Z) - qDKT: Question-centric Deep Knowledge Tracing [29.431121650577396]
We introduce qDKT, a variant of DKT that models every learner's success probability on individual questions over time.
qDKT incorporates graph Laplacian regularization to smooth predictions under each skill.
Experiments on several real-world datasets show that qDKT achieves state-of-art performance on predicting learner outcomes.
arXiv Detail & Related papers (2020-05-25T23:43:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.