Towards an Appropriate Query, Key, and Value Computation for Knowledge
Tracing
- URL: http://arxiv.org/abs/2002.07033v5
- Date: Mon, 1 Feb 2021 02:42:50 GMT
- Title: Towards an Appropriate Query, Key, and Value Computation for Knowledge
Tracing
- Authors: Youngduck Choi, Youngnam Lee, Junghyun Cho, Jineon Baek, Byungsoo Kim,
Yeongmin Cha, Dongmin Shin, Chan Bae, Jaewe Heo
- Abstract summary: We propose a novel Transformer based model for knowledge tracing, SAINT: Separated Self-AttentIve Neural Knowledge Tracing.
SAINT has an encoder-decoder structure where exercise and response embedding sequence separately enter the encoder and the decoder respectively.
This is the first work to suggest an encoder-decoder model for knowledge tracing that applies deep self-attentive layers to exercises and responses separately.
- Score: 2.1541440354538564
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Knowledge tracing, the act of modeling a student's knowledge through learning
activities, is an extensively studied problem in the field of computer-aided
education. Although models with attention mechanism have outperformed
traditional approaches such as Bayesian knowledge tracing and collaborative
filtering, they share two limitations. Firstly, the models rely on shallow
attention layers and fail to capture complex relations among exercises and
responses over time. Secondly, different combinations of queries, keys and
values for the self-attention layer for knowledge tracing were not extensively
explored. Usual practice of using exercises and interactions (exercise-response
pairs) as queries and keys/values respectively lacks empirical support. In this
paper, we propose a novel Transformer based model for knowledge tracing, SAINT:
Separated Self-AttentIve Neural Knowledge Tracing. SAINT has an encoder-decoder
structure where exercise and response embedding sequence separately enter the
encoder and the decoder respectively, which allows to stack attention layers
multiple times. To the best of our knowledge, this is the first work to suggest
an encoder-decoder model for knowledge tracing that applies deep self-attentive
layers to exercises and responses separately. The empirical evaluations on a
large-scale knowledge tracing dataset show that SAINT achieves the
state-of-the-art performance in knowledge tracing with the improvement of AUC
by 1.8% compared to the current state-of-the-art models.
Related papers
- KBAlign: Efficient Self Adaptation on Specific Knowledge Bases [75.78948575957081]
Large language models (LLMs) usually rely on retrieval-augmented generation to exploit knowledge materials in an instant manner.
We propose KBAlign, an approach designed for efficient adaptation to downstream tasks involving knowledge bases.
Our method utilizes iterative training with self-annotated data such as Q&A pairs and revision suggestions, enabling the model to grasp the knowledge content efficiently.
arXiv Detail & Related papers (2024-11-22T08:21:03Z) - Coupling Machine Learning with Ontology for Robotics Applications [0.0]
The lack of availability of prior knowledge in dynamic scenarios is without doubt a major barrier for scalable machine intelligence.
My view of the interaction between the two tiers intelligence is based on the idea that when knowledge is not readily available at the knowledge base tier, more knowledge can be extracted from the other tier.
arXiv Detail & Related papers (2024-06-08T23:38:03Z) - Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - UNTER: A Unified Knowledge Interface for Enhancing Pre-trained Language
Models [100.4659557650775]
We propose a UNified knowledge inTERface, UNTER, to provide a unified perspective to exploit both structured knowledge and unstructured knowledge.
With both forms of knowledge injected, UNTER gains continuous improvements on a series of knowledge-driven NLP tasks.
arXiv Detail & Related papers (2023-05-02T17:33:28Z) - A Unified End-to-End Retriever-Reader Framework for Knowledge-based VQA [67.75989848202343]
This paper presents a unified end-to-end retriever-reader framework towards knowledge-based VQA.
We shed light on the multi-modal implicit knowledge from vision-language pre-training models to mine its potential in knowledge reasoning.
Our scheme is able to not only provide guidance for knowledge retrieval, but also drop these instances potentially error-prone towards question answering.
arXiv Detail & Related papers (2022-06-30T02:35:04Z) - elBERto: Self-supervised Commonsense Learning for Question Answering [131.51059870970616]
We propose a Self-supervised Bidirectional Representation Learning of Commonsense framework, which is compatible with off-the-shelf QA model architectures.
The framework comprises five self-supervised tasks to force the model to fully exploit the additional training signals from contexts containing rich commonsense.
elBERto achieves substantial improvements on out-of-paragraph and no-effect questions where simple lexical similarity comparison does not help.
arXiv Detail & Related papers (2022-03-17T16:23:45Z) - Distilling Holistic Knowledge with Graph Neural Networks [37.86539695906857]
Knowledge Distillation (KD) aims at transferring knowledge from a larger well-optimized teacher network to a smaller learnable student network.
Existing KD methods have mainly considered two types of knowledge, namely the individual knowledge and the relational knowledge.
We propose to distill the novel holistic knowledge based on an attributed graph constructed among instances.
arXiv Detail & Related papers (2021-08-12T02:47:59Z) - ConCAD: Contrastive Learning-based Cross Attention for Sleep Apnea
Detection [16.938983046369263]
We propose a contrastive learning-based cross attention framework for sleep apnea detection (named ConCAD)
Our proposed framework can be easily integrated into standard deep learning models to utilize expert knowledge and contrastive learning to boost performance.
arXiv Detail & Related papers (2021-05-07T02:38:56Z) - Towards a Universal Continuous Knowledge Base [49.95342223987143]
We propose a method for building a continuous knowledge base that can store knowledge imported from multiple neural networks.
Experiments on text classification show promising results.
We import the knowledge from multiple models to the knowledge base, from which the fused knowledge is exported back to a single model.
arXiv Detail & Related papers (2020-12-25T12:27:44Z) - Deep Knowledge Tracing with Learning Curves [0.9088303226909278]
We propose a Convolution-Augmented Knowledge Tracing (CAKT) model in this paper.
The model employs three-dimensional convolutional neural networks to explicitly learn a student's recent experience on applying the same knowledge concept with that in the next question.
CAKT achieves the new state-of-the-art performance in predicting students' responses compared with existing models.
arXiv Detail & Related papers (2020-07-26T15:24:51Z) - HGKT: Introducing Hierarchical Exercise Graph for Knowledge Tracing [19.416373111152613]
We propose a hierarchical graph knowledge tracing model called HGKT to explore the latent hierarchical relations between exercises.
Specifically, we introduce the concept of problem schema to construct a hierarchical exercise graph that could model the exercise learning dependencies.
In the testing stage, we present a K&S diagnosis matrix that could trace the transition of mastery of knowledge and problem schema.
arXiv Detail & Related papers (2020-06-13T07:09:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.