TLSQKT: A Question-Aware Dual-Channel Transformer for Literacy Tracing from Learning Sequences
- URL: http://arxiv.org/abs/2510.22488v1
- Date: Sun, 26 Oct 2025 02:07:45 GMT
- Title: TLSQKT: A Question-Aware Dual-Channel Transformer for Literacy Tracing from Learning Sequences
- Authors: Zhifeng Wang, Yaowei Dong, Chunyan Zeng,
- Abstract summary: Knowledge tracing supports personalized learning by modeling how students' knowledge states evolve over time.<n>We instantiate this paradigm with a Transformer-based model, TLSQKT (Transformer for Learning Sequences with Question-Aware Knowledge Tracing)<n> TLSQKT employs a dual-channel design that jointly encodes student responses and item semantics, while question-aware interaction and self-attention capture long-range dependencies in learners' evolving states.
- Score: 4.119581464024065
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Knowledge tracing (KT) supports personalized learning by modeling how students' knowledge states evolve over time. However, most KT models emphasize mastery of discrete knowledge components, limiting their ability to characterize broader literacy development. We reframe the task as Literacy Tracing (LT), which models the growth of higher-order cognitive abilities and literacy from learners' interaction sequences, and we instantiate this paradigm with a Transformer-based model, TLSQKT (Transformer for Learning Sequences with Question-Aware Knowledge Tracing). TLSQKT employs a dual-channel design that jointly encodes student responses and item semantics, while question-aware interaction and self-attention capture long-range dependencies in learners' evolving states. Experiments on three real-world datasets - one public benchmark, one private knowledge-component dataset, and one private literacy dataset - show that TLSQKT consistently outperforms strong KT baselines on literacy-oriented metrics and reveals interpretable developmental trajectories of learners' literacy. Transfer experiments further indicate that knowledge-tracing signals can be leveraged for literacy tracing, offering a practical route when dedicated literacy labels are limited. These findings position literacy tracing as a scalable component of intelligent educational systems and lay the groundwork for literacy evaluation in future large-scale educational models.
Related papers
- Towards LLM-Empowered Knowledge Tracing via LLM-Student Hierarchical Behavior Alignment in Hyperbolic Space [24.868649493405528]
Knowledge Tracing (KT) diagnoses students' concept mastery through continuous learning state monitoring in education.<n>Existing methods rely on ID-based sequences or shallow textual features.<n>This paper proposes a Large Language Model Hyperbolic Aligned Knowledge Tracing framework.
arXiv Detail & Related papers (2026-02-26T11:17:31Z) - Towards Improving Interpretability of Language Model Generation through a Structured Knowledge Discovery Approach [33.17711262799183]
We develop a task-agnostic structured knowledge hunter for knowledge-enhanced text generation tasks.<n>Our model achieves high interpretability, enabling users to comprehend the model output generation process.<n>We empirically demonstrate the effectiveness of our model in both internal knowledge-enhanced table-to-text generation on the RotoWireFG dataset and external knowledge-enhanced dialogue response generation on the KdConv dataset.
arXiv Detail & Related papers (2025-11-28T16:43:46Z) - Next Token Knowledge Tracing: Exploiting Pretrained LLM Representations to Decode Student Behaviour [5.32438871812364]
The Knowledge Tracing task aims to predict how students will respond to educational questions in learning environments.<n>Existing KT models typically use response correctness along with metadata like skill tags and timestamps, often overlooking the question text.<n>We propose Next Token Knowledge Tracing (NTKT), a novel approach that reframes KT as a next-token prediction task using pretrained Large Language Models.
arXiv Detail & Related papers (2025-11-04T14:20:56Z) - TRAIL: Joint Inference and Refinement of Knowledge Graphs with Large Language Models [5.678291291711662]
TRAIL is a novel, unified framework for Thinking, Reasoning, And Incremental Learning.<n>It couples joint inference and dynamic KG refinement with large language models.<n>Extensive experiments on multiple benchmarks demonstrate that TRAIL outperforms existing KG-augmented and retrieval-augmented LLM baselines by 3% to 13%.
arXiv Detail & Related papers (2025-08-06T14:25:05Z) - LLM4CD: Leveraging Large Language Models for Open-World Knowledge Augmented Cognitive Diagnosis [56.50378080174923]
We propose LLM4CD, which Leverages Large Language Models for Open-World Knowledge Augmented Cognitive Diagnosis.<n>Our method utilizes the open-world knowledge of LLMs to construct cognitively expressive textual representations, which are encoded to introduce rich semantic information into the CD task.<n>This approach substitutes traditional ID embeddings with semantic representations, enabling the model to accommodate new students and exercises with open-world knowledge and address the cold-start problem.
arXiv Detail & Related papers (2025-05-14T14:48:00Z) - Contrastive Cross-Course Knowledge Tracing via Concept Graph Guided Knowledge Transfer [12.34590941832835]
We propose TransKT, a contrastive cross-course knowledge tracing method.<n>It builds on concept graph guided knowledge transfer to model the relationships between learning behaviors across different courses.<n>TransKT employs a contrastive objective that aligns single-course and cross-course knowledge states.
arXiv Detail & Related papers (2025-05-14T10:38:30Z) - Enhancing LLM Knowledge Learning through Generalization [73.16975077770765]
We show that an LLM's ability to continually predict the same factual knowledge tokens given diverse paraphrased contexts is positively correlated with its capacity to extract that knowledge via question-answering.<n>We propose two strategies to enhance LLMs' ability to predict the same knowledge tokens given varied contexts, thereby enhancing knowledge acquisition.
arXiv Detail & Related papers (2025-03-05T17:56:20Z) - A Systematic Review of Knowledge Tracing and Large Language Models in Education: Opportunities, Issues, and Future Research [0.0]
Large Language Models (LLMs) are pre-trained on vast natural language datasets.<n>This systematic review explores the intersections, opportunities, and challenges of combining KT models and LLMs in educational contexts.
arXiv Detail & Related papers (2024-12-12T13:00:50Z) - KaLM: Knowledge-aligned Autoregressive Language Modeling via Dual-view Knowledge Graph Contrastive Learning [74.21524111840652]
This paper proposes textbfKaLM, a textitKnowledge-aligned Language Modeling approach.<n>It fine-tunes autoregressive large language models to align with KG knowledge via the joint objective of explicit knowledge alignment and implicit knowledge alignment.<n> Notably, our method achieves a significant performance boost in evaluations of knowledge-driven tasks.
arXiv Detail & Related papers (2024-12-06T11:08:24Z) - Tabular Transfer Learning via Prompting LLMs [52.96022335067357]
We propose a novel framework, Prompt to Transfer (P2T), that utilizes unlabeled (or heterogeneous) source data with large language models (LLMs)
P2T identifies a column feature in a source dataset that is strongly correlated with a target task feature to create examples relevant to the target task, thus creating pseudo-demonstrations for prompts.
arXiv Detail & Related papers (2024-08-09T11:30:52Z) - Knowledge Graph-Enhanced Large Language Models via Path Selection [58.228392005755026]
Large Language Models (LLMs) have shown unprecedented performance in various real-world applications.
LLMs are known to generate factually inaccurate outputs, a.k.a. the hallucination problem.
We propose a principled framework KELP with three stages to handle the above problems.
arXiv Detail & Related papers (2024-06-19T21:45:20Z) - Leveraging Pedagogical Theories to Understand Student Learning Process with Graph-based Reasonable Knowledge Tracing [11.082908318943248]
We introduce GRKT, a graph-based reasonable knowledge tracing method to address these issues.
We propose a fine-grained and psychological three-stage modeling process as knowledge retrieval, memory strengthening, and knowledge learning/forgetting.
arXiv Detail & Related papers (2024-06-07T10:14:30Z) - Predictive, scalable and interpretable knowledge tracing on structured domains [6.860460230412773]
PSI-KT is a hierarchical generative approach that explicitly models how both individual cognitive traits and the prerequisite structure of knowledge influence learning dynamics.
PSI-KT targets the real-world need for efficient personalization even with a growing body of learners and learning histories.
arXiv Detail & Related papers (2024-03-19T22:19:29Z) - Knowledgeable Salient Span Mask for Enhancing Language Models as
Knowledge Base [51.55027623439027]
We develop two solutions to help the model learn more knowledge from unstructured text in a fully self-supervised manner.
To our best knowledge, we are the first to explore fully self-supervised learning of knowledge in continual pre-training.
arXiv Detail & Related papers (2022-04-17T12:33:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.