Leveraging Deep Reinforcement Learning for Metacognitive Interventions
across Intelligent Tutoring Systems
- URL: http://arxiv.org/abs/2304.09821v1
- Date: Mon, 17 Apr 2023 12:10:50 GMT
- Title: Leveraging Deep Reinforcement Learning for Metacognitive Interventions
across Intelligent Tutoring Systems
- Authors: Mark Abdelshiheed, John Wesley Hostetter, Tiffany Barnes, Min Chi
- Abstract summary: This work compares two approaches to provide metacognitive interventions across Intelligent Tutoring Systems (ITSs)
In two consecutive semesters, we conducted two classroom experiments: Exp. 1 used a classic artificial intelligence approach to classify students into different metacognitive groups and provide static interventions based on their classified groups.
In Exp. 2, we leveraged Deep Reinforcement Learning (DRL) to provide adaptive interventions that consider the dynamic changes in the student's metacognitive levels.
- Score: 7.253181280137071
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work compares two approaches to provide metacognitive interventions and
their impact on preparing students for future learning across Intelligent
Tutoring Systems (ITSs). In two consecutive semesters, we conducted two
classroom experiments: Exp. 1 used a classic artificial intelligence approach
to classify students into different metacognitive groups and provide static
interventions based on their classified groups. In Exp. 2, we leveraged Deep
Reinforcement Learning (DRL) to provide adaptive interventions that consider
the dynamic changes in the student's metacognitive levels. In both experiments,
students received these interventions that taught how and when to use a
backward-chaining (BC) strategy on a logic tutor that supports a default
forward-chaining strategy. Six weeks later, we trained students on a
probability tutor that only supports BC without interventions. Our results show
that adaptive DRL-based interventions closed the metacognitive skills gap
between students. In contrast, static classifier-based interventions only
benefited a subset of students who knew how to use BC in advance. Additionally,
our DRL agent prepared the experimental students for future learning by
significantly surpassing their control peers on both ITSs.
Related papers
- Chaos-based reinforcement learning with TD3 [3.04503073434724]
Chaos-based reinforcement learning (CBRL) is a method in which the agent's internal chaotic dynamics drives exploration.
This study introduced Twin Delayed Deep Deterministic Policy Gradients (TD3), which is one of the state-of-the-art deep reinforcement learning algorithms.
arXiv Detail & Related papers (2024-05-15T04:47:31Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - Demonstration-free Autonomous Reinforcement Learning via Implicit and
Bidirectional Curriculum [22.32327908453603]
We propose a demonstration-free reinforcement learning algorithm via Implicit and Bi-directional Curriculum (IBC)
With an auxiliary agent that is conditionally activated upon learning progress and a bidirectional goal curriculum based on optimal transport, our method outperforms previous methods.
arXiv Detail & Related papers (2023-05-17T04:31:36Z) - Bridging Declarative, Procedural, and Conditional Metacognitive
Knowledge Gap Using Deep Reinforcement Learning [7.253181280137071]
In deductive domains, three metacognitive knowledge types in ascending order are declarative, procedural, and conditional learning.
This work leverages Deep Reinforcement Learning (DRL) in providing adaptive metacognitive interventions to bridge the gap between the three knowledge types.
Our results show that on both ITSs, DRL bridged the metacognitive knowledge gap between students and significantly improved their learning performance over their control peers.
arXiv Detail & Related papers (2023-04-23T20:07:07Z) - Mixing Backward- with Forward-Chaining for Metacognitive Skill
Acquisition and Transfer [9.702049806385011]
Students were trained on a logic tutor that supports a default forward-chaining (FC) and a backward-chaining (BC) strategy.
We investigated the impact of mixing BC with FC on teaching strategy- and time-awareness for nonStrTime students.
arXiv Detail & Related papers (2023-03-18T16:44:10Z) - Distantly-Supervised Named Entity Recognition with Adaptive Teacher
Learning and Fine-grained Student Ensemble [56.705249154629264]
Self-training teacher-student frameworks are proposed to improve the robustness of NER models.
In this paper, we propose an adaptive teacher learning comprised of two teacher-student networks.
Fine-grained student ensemble updates each fragment of the teacher model with a temporal moving average of the corresponding fragment of the student, which enhances consistent predictions on each model fragment against noise.
arXiv Detail & Related papers (2022-12-13T12:14:09Z) - Enhancing Adversarial Training with Feature Separability [52.39305978984573]
We introduce a new concept of adversarial training graph (ATG) with which the proposed adversarial training with feature separability (ATFS) enables to boost the intra-class feature similarity and increase inter-class feature variance.
Through comprehensive experiments, we demonstrate that the proposed ATFS framework significantly improves both clean and robust performance.
arXiv Detail & Related papers (2022-05-02T04:04:23Z) - Autonomous Reinforcement Learning: Formalism and Benchmarking [106.25788536376007]
Real-world embodied learning, such as that performed by humans and animals, is situated in a continual, non-episodic world.
Common benchmark tasks in RL are episodic, with the environment resetting between trials to provide the agent with multiple attempts.
This discrepancy presents a major challenge when attempting to take RL algorithms developed for episodic simulated environments and run them on real-world platforms.
arXiv Detail & Related papers (2021-12-17T16:28:06Z) - Persistent Reinforcement Learning via Subgoal Curricula [114.83989499740193]
Value-accelerated Persistent Reinforcement Learning (VaPRL) generates a curriculum of initial states.
VaPRL reduces the interventions required by three orders of magnitude compared to episodic reinforcement learning.
arXiv Detail & Related papers (2021-07-27T16:39:45Z) - LANA: Towards Personalized Deep Knowledge Tracing Through
Distinguishable Interactive Sequences [21.67751919579854]
We propose Leveled Attentive KNowledge TrAcing (LANA) to predict students' responses to future questions.
It uses a novel student-related features extractor (SRFE) to distill students' unique inherent properties from their respective interactive sequences.
With pivot module reconstructed the decoder for individual students and leveled learning specialized encoders for groups, personalized DKT was achieved.
arXiv Detail & Related papers (2021-04-21T02:57:42Z) - Transfer Heterogeneous Knowledge Among Peer-to-Peer Teammates: A Model
Distillation Approach [55.83558520598304]
We propose a brand new solution to reuse experiences and transfer value functions among multiple students via model distillation.
We also describe how to design an efficient communication protocol to exploit heterogeneous knowledge.
Our proposed framework, namely Learning and Teaching Categorical Reinforcement, shows promising performance on stabilizing and accelerating learning progress.
arXiv Detail & Related papers (2020-02-06T11:31:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.