Ruffle&Riley: Insights from Designing and Evaluating a Large Language Model-Based Conversational Tutoring System
- URL: http://arxiv.org/abs/2404.17460v1
- Date: Fri, 26 Apr 2024 14:57:55 GMT
- Title: Ruffle&Riley: Insights from Designing and Evaluating a Large Language Model-Based Conversational Tutoring System
- Authors: Robin Schmucker, Meng Xia, Amos Azaria, Tom Mitchell,
- Abstract summary: Conversational tutoring systems (CTSs) offer learning experiences through interactions based on natural language.
We discuss and evaluate a novel type of CTS that leverages recent advances in large language models (LLMs) in two ways.
The system enables AI-assisted content authoring by inducing an easily editable tutoring script automatically from a lesson text.
- Score: 21.139850269835858
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conversational tutoring systems (CTSs) offer learning experiences through interactions based on natural language. They are recognized for promoting cognitive engagement and improving learning outcomes, especially in reasoning tasks. Nonetheless, the cost associated with authoring CTS content is a major obstacle to widespread adoption and to research on effective instructional design. In this paper, we discuss and evaluate a novel type of CTS that leverages recent advances in large language models (LLMs) in two ways: First, the system enables AI-assisted content authoring by inducing an easily editable tutoring script automatically from a lesson text. Second, the system automates the script orchestration in a learning-by-teaching format via two LLM-based agents (Ruffle&Riley) acting as a student and a professor. The system allows for free-form conversations that follow the ITS-typical inner and outer loop structure. We evaluate Ruffle&Riley's ability to support biology lessons in two between-subject online user studies (N = 200) comparing the system to simpler QA chatbots and reading activity. Analyzing system usage patterns, pre/post-test scores and user experience surveys, we find that Ruffle&Riley users report high levels of engagement, understanding and perceive the offered support as helpful. Even though Ruffle&Riley users require more time to complete the activity, we did not find significant differences in short-term learning gains over the reading activity. Our system architecture and user study provide various insights for designers of future CTSs. We further open-source our system to support ongoing research on effective instructional design of LLM-based learning technologies.
Related papers
- Let Students Take the Wheel: Introducing Post-Quantum Cryptography with Active Learning [4.804847392457553]
Post-quantum cryptography (PQC) has been identified as the solution to secure existing software systems.
This research proposes a novel active learning approach and assesses the best practices for teaching PQC to undergraduate and graduate students.
arXiv Detail & Related papers (2024-10-17T01:52:03Z) - Exploring Knowledge Tracing in Tutor-Student Dialogues [53.52699766206808]
We present a first attempt at performing knowledge tracing (KT) in tutor-student dialogues.
We propose methods to identify the knowledge components/skills involved in each dialogue turn.
We then apply a range of KT methods on the resulting labeled data to track student knowledge levels over an entire dialogue.
arXiv Detail & Related papers (2024-09-24T22:31:39Z) - Pronunciation Assessment with Multi-modal Large Language Models [10.35401596425946]
We propose a scoring system based on large language models (LLMs)
The speech encoder first maps the learner's speech into contextual features.
The adapter layer then transforms these features to align with the text embedding in latent space.
arXiv Detail & Related papers (2024-07-12T12:16:14Z) - SPL: A Socratic Playground for Learning Powered by Large Language Model [5.383689446227398]
Socratic Playground for Learning (SPL) is a dialogue-based ITS powered by the GPT-4 model.
SPL aims to enhance personalized and adaptive learning experiences tailored to individual needs.
arXiv Detail & Related papers (2024-06-20T01:18:52Z) - DIALIGHT: Lightweight Multilingual Development and Evaluation of
Task-Oriented Dialogue Systems with Large Language Models [76.79929883963275]
DIALIGHT is a toolkit for developing and evaluating multilingual Task-Oriented Dialogue (ToD) systems.
It features a secure, user-friendly web interface for fine-grained human evaluation at both local utterance level and global dialogue level.
Our evaluations reveal that while PLM fine-tuning leads to higher accuracy and coherence, LLM-based systems excel in producing diverse and likeable responses.
arXiv Detail & Related papers (2024-01-04T11:27:48Z) - Ruffle&Riley: Towards the Automated Induction of Conversational Tutoring
Systems [23.093767743306973]
Conversational tutoring systems (CTSs) offer learning experiences driven by natural language interaction.
We introduce a novel type of CTS that leverages the recent advances in large language models (LLMs) in two ways.
Ruffle&Riley allows a free-form conversation that follows the ITS-typical inner and outer loop structure.
arXiv Detail & Related papers (2023-09-26T23:27:06Z) - Empowering Private Tutoring by Chaining Large Language Models [87.76985829144834]
This work explores the development of a full-fledged intelligent tutoring system powered by state-of-the-art large language models (LLMs)
The system is into three inter-connected core processes-interaction, reflection, and reaction.
Each process is implemented by chaining LLM-powered tools along with dynamically updated memory modules.
arXiv Detail & Related papers (2023-09-15T02:42:03Z) - AI Chatbots as Multi-Role Pedagogical Agents: Transforming Engagement in
CS Education [8.898863361318817]
We develop, implement, and evaluate a novel learning environment enriched with four distinct chatbots.
These roles cater to the three innate psychological needs of learners - competence, autonomy, and relatedness.
The system embraces an inquiry-based learning paradigm, encouraging students to ask questions, seek solutions, and explore their curiosities.
arXiv Detail & Related papers (2023-08-08T02:13:44Z) - An Adversarial Learning based Multi-Step Spoken Language Understanding
System through Human-Computer Interaction [70.25183730482915]
We introduce a novel multi-step spoken language understanding system based on adversarial learning.
We demonstrate that the new system can improve parsing performance by at least $2.5%$ in terms of F1.
arXiv Detail & Related papers (2021-06-06T03:46:53Z) - Continual Learning in Task-Oriented Dialogue Systems [49.35627673523519]
Continual learning in task-oriented dialogue systems can allow us to add new domains and functionalities through time without incurring the high cost of a whole system retraining.
We propose a continual learning benchmark for task-oriented dialogue systems with 37 domains to be learned continuously in four settings.
arXiv Detail & Related papers (2020-12-31T08:44:25Z) - MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems [75.43457658815943]
We propose Minimalist Transfer Learning (MinTL) to simplify the system design process of task-oriented dialogue systems.
MinTL is a simple yet effective transfer learning framework, which allows us to plug-and-play pre-trained seq2seq models.
We instantiate our learning framework with two pre-trained backbones: T5 and BART, and evaluate them on MultiWOZ.
arXiv Detail & Related papers (2020-09-25T02:19:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.