Playsemble: Learning Low-Level Programming Through Interactive Games
- URL: http://arxiv.org/abs/2602.20167v1
- Date: Mon, 09 Feb 2026 06:31:44 GMT
- Title: Playsemble: Learning Low-Level Programming Through Interactive Games
- Authors: Elliott Wen, Paul Denny, Andrew Luxton-Reilly, Sean Ma, Bruce Sham, Chenye Ni, Jun Seo, Yu Yang,
- Abstract summary: Playsemble is a gamified learning system that transforms assembly instructions into interactive, game-like tasks.<n>Our findings suggest that Playsemble promotes active experimentation, sustained engagement, and deeper conceptual understanding.
- Score: 6.174988590679205
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Teaching assembly programming is a fundamental component of undergraduate computer science education, yet many students struggle with its abstract and low-level concepts. Existing learning tools, such as simulators and visualisers, support understanding by exposing machine states. However, they often limit students to passive observation and provide few opportunities for meaningful interaction. To address these limitations, we introduce Playsemble, a gamified learning system that transforms assembly instructions into interactive, game-like tasks in which students control Pac-Man to collect items, avoid ghosts, and reach targets. Playsemble integrates a code editor, a CPU emulator, and visual debugging tools within a browser-based environment, allowing students to work offline without installation or configuration. It also provides immediate formative feedback enhanced by large language models. We deployed Playsemble in an undergraduate computer architecture course with 107 students. The course featured a sequence of assignments of increasing complexity, covering core concepts such as register and memory manipulation, control structures including loops and conditionals, and arithmetic operations. Our findings suggest that Playsemble promotes active experimentation, sustained engagement, and deeper conceptual understanding through meaningful game-based learning experiences.
Related papers
- Bridging Pedagogy and Play: Introducing a Language Mapping Interface for Human-AI Co-Creation in Educational Game Design [15.064660952504129]
Existing authoring environments do not eliminate the underlying challenges of educational game design.<n>We designed a controlled natural language framework-based web tool that positions language as the primary interface for LLM-assisted educational game design.<n>We argue that, by making pedagogical intent explicit and editable in the interface, the tool has the potential to lower design barriers for non-expert designers, preserves human agency in critical decisions, and enables alignment and reflections between pedagogy and gameplay during and after co-creation.
arXiv Detail & Related papers (2026-03-04T02:08:32Z) - Sojourner under Sabotage: A Serious Testing and Debugging Game [9.856068089918555]
Sojourner under Sabotage is a browser-based serious game that blends education with an immersive storyline.<n>A study with 79 students demonstrates that the game is a powerful tool for enhancing motivation, engagement, and skill development.
arXiv Detail & Related papers (2025-04-27T16:05:10Z) - Playpen: An Environment for Exploring Learning Through Conversational Interaction [84.0413820245725]
We investigate whether Dialogue Games can also serve as a source of feedback signals for learning.<n>We introduce Playpen, an environment for off- and online learning through Dialogue Game self-play.<n>We find that imitation learning through SFT improves performance on unseen instances, but negatively impacts other skills.
arXiv Detail & Related papers (2025-04-11T14:49:33Z) - Inclusive STEAM Education: A Framework for Teaching Cod-2 ing and Robotics to Students with Visually Impairment Using 3 Advanced Computer Vision [0.0]
This paper presents a framework that leverages pre-constructed robots and algorithms, such as maze-solving techniques, within an accessible learning environment.<n>The proposed system employs Contrastive Language-Image Pre-training (CLIP) to process global camera-captured maze layouts.<n>Students issue verbal commands, which are refined through CLIP, while robot-mounted stereo cameras provide real-time data processed via Simultaneous SLAM.
arXiv Detail & Related papers (2025-03-06T17:15:12Z) - ClawMachine: Learning to Fetch Visual Tokens for Referential Comprehension [71.03445074045092]
We propose ClawMachine, offering a new methodology that explicitly notates each entity using token collectives groups of visual tokens.<n>Our method unifies the prompt and answer of visual referential tasks without using additional syntax.<n>ClawMachine achieves superior performance on scene-level and referential understanding tasks with higher efficiency.
arXiv Detail & Related papers (2024-06-17T08:39:16Z) - Learning of Generalizable and Interpretable Knowledge in Grid-Based
Reinforcement Learning Environments [5.217870815854702]
We propose using program synthesis to imitate reinforcement learning policies.
We adapt the state-of-the-art program synthesis system DreamCoder for learning concepts in grid-based environments.
arXiv Detail & Related papers (2023-09-07T11:46:57Z) - Embodied Concept Learner: Self-supervised Learning of Concepts and
Mapping through Instruction Following [101.55727845195969]
We propose Embodied Learner Concept (ECL) in an interactive 3D environment.
A robot agent can ground visual concepts, build semantic maps and plan actions to complete tasks.
ECL is fully transparent and step-by-step interpretable in long-term planning.
arXiv Detail & Related papers (2023-04-07T17:59:34Z) - Language-Driven Representation Learning for Robotics [115.93273609767145]
Recent work in visual representation learning for robotics demonstrates the viability of learning from large video datasets of humans performing everyday tasks.
We introduce a framework for language-driven representation learning from human videos and captions.
We find that Voltron's language-driven learning outperform the prior-of-the-art, especially on targeted problems requiring higher-level control.
arXiv Detail & Related papers (2023-02-24T17:29:31Z) - Orbit: A Unified Simulation Framework for Interactive Robot Learning
Environments [38.23943905182543]
We present Orbit, a unified and modular framework for robot learning powered by NVIDIA Isaac Sim.
It offers a modular design to create robotic environments with photo-realistic scenes and high-fidelity rigid and deformable body simulation.
We aim to support various research areas, including representation learning, reinforcement learning, imitation learning, and task and motion planning.
arXiv Detail & Related papers (2023-01-10T20:19:17Z) - Deep Reinforcement Learning with Stacked Hierarchical Attention for
Text-based Games [64.11746320061965]
We study reinforcement learning for text-based games, which are interactive simulations in the context of natural language.
We aim to conduct explicit reasoning with knowledge graphs for decision making, so that the actions of an agent are generated and supported by an interpretable inference procedure.
We extensively evaluate our method on a number of man-made benchmark games, and the experimental results demonstrate that our method performs better than existing text-based agents.
arXiv Detail & Related papers (2020-10-22T12:40:22Z) - Learning to Simulate Dynamic Environments with GameGAN [109.25308647431952]
In this paper, we aim to learn a simulator by simply watching an agent interact with an environment.
We introduce GameGAN, a generative model that learns to visually imitate a desired game by ingesting screenplay and keyboard actions during training.
arXiv Detail & Related papers (2020-05-25T14:10:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.