Embracing Imperfection: Simulating Students with Diverse Cognitive Levels Using LLM-based Agents
- URL: http://arxiv.org/abs/2505.19997v1
- Date: Mon, 26 May 2025 13:48:49 GMT
- Title: Embracing Imperfection: Simulating Students with Diverse Cognitive Levels Using LLM-based Agents
- Authors: Tao Wu, Jingyuan Chen, Wang Lin, Mengze Li, Yumeng Zhu, Ang Li, Kun Kuang, Fei Wu,
- Abstract summary: Large language models (LLMs) are revolutionizing education, with LLM-based agents playing a key role in simulating student behavior.<n>A major challenge in student simulation is modeling the diverse learning patterns of students at various cognitive levels.
- Score: 36.704574105201864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large language models (LLMs) are revolutionizing education, with LLM-based agents playing a key role in simulating student behavior. A major challenge in student simulation is modeling the diverse learning patterns of students at various cognitive levels. However, current LLMs, typically trained as ``helpful assistants'', target at generating perfect responses. As a result, they struggle to simulate students with diverse cognitive abilities, as they often produce overly advanced answers, missing the natural imperfections that characterize student learning and resulting in unrealistic simulations. To address this issue, we propose a training-free framework for student simulation. We begin by constructing a cognitive prototype for each student using a knowledge graph, which captures their understanding of concepts from past learning records. This prototype is then mapped to new tasks to predict student performance. Next, we simulate student solutions based on these predictions and iteratively refine them using a beam search method to better replicate realistic mistakes. To validate our approach, we construct the \texttt{Student\_100} dataset, consisting of $100$ students working on Python programming and $5,000$ learning records. Experimental results show that our method consistently outperforms baseline models, achieving $100\%$ improvement in simulation accuracy.
Related papers
- ParaStudent: Generating and Evaluating Realistic Student Code by Teaching LLMs to Struggle [24.691302820912888]
Large Language Models (LLMs) have shown strong performance on programming tasks, but can they generate student-like code like real students?<n>We present ParaStudent, a systematic study of LLM-based "student-like" code generation in an introductory programming course setting.
arXiv Detail & Related papers (2025-07-16T23:12:14Z) - Classroom Simulacra: Building Contextual Student Generative Agents in Online Education for Learning Behavioral Simulation [10.209326669619273]
We run a 6-week education workshop from N = 60 students to collect fine-grained data using a custom built online education system.<n>We propose a transferable iterative reflection (TIR) module that augments both prompting-based and finetuning-based large language models.
arXiv Detail & Related papers (2025-02-04T23:42:52Z) - LLM-based Cognitive Models of Students with Misconceptions [55.29525439159345]
This paper investigates whether Large Language Models (LLMs) can be instruction-tuned to meet this dual requirement.
We introduce MalAlgoPy, a novel Python library that generates datasets reflecting authentic student solution patterns.
Our insights enhance our understanding of AI-based student models and pave the way for effective adaptive learning systems.
arXiv Detail & Related papers (2024-10-16T06:51:09Z) - Toward In-Context Teaching: Adapting Examples to Students' Misconceptions [54.82965010592045]
We introduce a suite of models and evaluation methods we call AdapT.
AToM is a new probabilistic model for adaptive teaching that jointly infers students' past beliefs and optimize for the correctness of future beliefs.
Our results highlight both the difficulty of the adaptive teaching task and the potential of learned adaptive models for solving it.
arXiv Detail & Related papers (2024-05-07T17:05:27Z) - EduAgent: Generative Student Agents in Learning [15.215078619481732]
Student simulation in online education is important to address dynamic learning behaviors of students with diverse backgrounds.
Existing simulation models based on deep learning usually need massive training data, lacking prior knowledge in educational contexts.
This work proposes EduAgent, a novel generative agent framework incorporating cognitive prior knowledge.
arXiv Detail & Related papers (2024-03-23T18:19:17Z) - Leveraging generative artificial intelligence to simulate student
learning behavior [13.171768256928509]
We explore the feasibility of using large language models (LLMs) to simulate student learning behaviors.
Unlike conventional machine learning based prediction, we leverage LLMs to instantiate virtual students with specific demographics.
Our objective is not merely to predict learning outcomes but to replicate learning behaviors and patterns of real students.
arXiv Detail & Related papers (2023-10-30T00:09:59Z) - User Behavior Simulation with Large Language Model based Agents [116.74368915420065]
We propose an LLM-based agent framework and design a sandbox environment to simulate real user behaviors.
Based on extensive experiments, we find that the simulated behaviors of our method are very close to the ones of real humans.
arXiv Detail & Related papers (2023-06-05T02:58:35Z) - Hindsight States: Blending Sim and Real Task Elements for Efficient
Reinforcement Learning [61.3506230781327]
In robotics, one approach to generate training data builds on simulations based on dynamics models derived from first principles.
Here, we leverage the imbalance in complexity of the dynamics to learn more sample-efficiently.
We validate our method on several challenging simulated tasks and demonstrate that it improves learning both alone and when combined with an existing hindsight algorithm.
arXiv Detail & Related papers (2023-03-03T21:55:04Z) - Generalisable Methods for Early Prediction in Interactive Simulations
for Education [5.725477071353353]
Classifying students' interaction data in the simulations based on their expected performance has the potential to enable adaptive guidance.
We first measure the students' conceptual understanding through their in-task performance.
Then, we suggest a novel type of features that, starting from clickstream data, encodes both the state of the simulation and the action performed by the student.
arXiv Detail & Related papers (2022-07-04T14:46:56Z) - Practical Imitation Learning in the Real World via Task Consistency Loss [18.827979446629296]
This paper introduces a self-supervised loss that encourages sim and real alignment both at the feature and action-prediction levels.
We achieve 80% success across ten seen and unseen scenes using only 16.2 hours of teleoperated demonstrations in sim and real.
arXiv Detail & Related papers (2022-02-03T21:43:06Z) - Visual Adversarial Imitation Learning using Variational Models [60.69745540036375]
Reward function specification remains a major impediment for learning behaviors through deep reinforcement learning.
Visual demonstrations of desired behaviors often presents an easier and more natural way to teach agents.
We develop a variational model-based adversarial imitation learning algorithm.
arXiv Detail & Related papers (2021-07-16T00:15:18Z) - Reactive Long Horizon Task Execution via Visual Skill and Precondition
Models [59.76233967614774]
We describe an approach for sim-to-real training that can accomplish unseen robotic tasks using models learned in simulation to ground components of a simple task planner.
We show an increase in success rate from 91.6% to 98% in simulation and from 10% to 80% success rate in the real-world as compared with naive baselines.
arXiv Detail & Related papers (2020-11-17T15:24:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.