Measuring Student Behavioral Engagement using Histogram of Actions
- URL: http://arxiv.org/abs/2307.09420v2
- Date: Thu, 15 May 2025 14:30:03 GMT
- Title: Measuring Student Behavioral Engagement using Histogram of Actions
- Authors: Ahmed Abdelkawy, Aly Farag, Islam Alkabbany, Asem Ali, Chris Foreman, Thomas Tretter, Nicholas Hindy,
- Abstract summary: The proposed approach recognizes student actions then predicts the student behavioral engagement level.<n>For student action recognition, we use human skeletons to model student postures and upper body movements.<n>The trained 3D-CNN model is used to recognize actions within every 2minute video segment.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In this paper, we propose a novel technique for measuring behavioral engagement through students' actions recognition. The proposed approach recognizes student actions then predicts the student behavioral engagement level. For student action recognition, we use human skeletons to model student postures and upper body movements. To learn the dynamics of student upper body, a 3D-CNN model is used. The trained 3D-CNN model is used to recognize actions within every 2minute video segment then these actions are used to build a histogram of actions which encodes the student actions and their frequencies. This histogram is utilized as an input to SVM classifier to classify whether the student is engaged or disengaged. To evaluate the proposed framework, we build a dataset consisting of 1414 2-minute video segments annotated with 13 actions and 112 video segments annotated with two engagement levels. Experimental results indicate that student actions can be recognized with top 1 accuracy 83.63% and the proposed framework can capture the average engagement of the class.
Related papers
- Context Matters: Peer-Aware Student Behavioral Engagement Measurement via VLM Action Parsing and LLM Sequence Classification [0.6103775976356991]
We propose a novel three-stage framework for video-based student engagement measurement.<n>First, we explore the few-shot adaptation of the vision-language model for student action recognition.<n>Second, we utilize the sliding temporal window technique to divide each student's 2-minute-long video into non-overlapping segments.<n>Third, we leverage the large language model to classify this entire sequence of actions, together with the classroom context, as belonging to an engaged or disengaged student.
arXiv Detail & Related papers (2026-01-10T02:39:24Z) - Informative Sample Selection Model for Skeleton-based Action Recognition with Limited Training Samples [51.59753385094941]
3D Action Recognition with Limited Training Samples, also known as semi-supervised 3D Action Recognition, has been proposed.<n>We reformulate Semi-supervised 3D action recognition via active learning from a novel perspective by casting it as a Markov Decision Process (MDP)<n>To enhance the representational capacity of the factors in the state-action pairs within our method, we project them from Euclidean space to hyperbolic space.
arXiv Detail & Related papers (2025-10-29T10:03:33Z) - Supervised Contrastive Learning for Ordinal Engagement Measurement [2.166000001057538]
Student engagement plays a crucial role in the successful delivery of educational programs.<n>This paper identifies two key challenges in this problem: class imbalance and incorporating order into engagement levels.<n>A novel approach to video-based student engagement measurement in virtual learning environments is proposed.
arXiv Detail & Related papers (2025-05-27T03:49:45Z) - Improving Question Embeddings with Cognitiv Representation Optimization for Knowledge Tracing [77.14348157016518]
The Knowledge Tracing (KT) aims to track changes in students' knowledge status and predict their future answers based on their historical answer records.
Current research on KT modeling focuses on predicting student' future performance based on existing, unupdated records of student learning interactions.
We propose a Cognitive Representation Optimization for Knowledge Tracing model, which utilizes a dynamic programming algorithm to optimize structure of cognitive representations.
arXiv Detail & Related papers (2025-04-05T09:32:03Z) - Long-term Human Participation Assessment In Collaborative Learning Environments Using Dynamic Scene Analysis [2.115993069505241]
The paper develops datasets and methods to assess student participation in real-life collaborative learning environments.
We formulate the problem of assessing student participation into two subproblems: (i) student group detection against strong background interference from other groups, and (ii) dynamic participant tracking within the group.
arXiv Detail & Related papers (2024-04-14T21:39:00Z) - Early Action Recognition with Action Prototypes [62.826125870298306]
We propose a novel model that learns a prototypical representation of the full action for each class.
We decompose the video into short clips, where a visual encoder extracts features from each clip independently.
Later, a decoder aggregates together in an online fashion features from all the clips for the final class prediction.
arXiv Detail & Related papers (2023-12-11T18:31:13Z) - Bag of States: A Non-sequential Approach to Video-based Engagement
Measurement [7.864500429933145]
Students' behavioral and emotional states need to be analyzed at fine-grained time scales in order to measure their level of engagement.
Many existing approaches have developed sequential andtemporal models, such as recurrent neural networks, temporal convolutional networks, and three-dimensional convolutional neural networks, for measuring student engagement from videos.
We develop bag-of-words-based models in which only occurrence of behavioral and emotional states of students is modeled and analyzed and not the order in which they occur.
arXiv Detail & Related papers (2023-01-17T07:12:34Z) - Learning Action-Effect Dynamics from Pairs of Scene-graphs [50.72283841720014]
We propose a novel method that leverages scene-graph representation of images to reason about the effects of actions described in natural language.
Our proposed approach is effective in terms of performance, data efficiency, and generalization capability compared to existing models.
arXiv Detail & Related papers (2022-12-07T03:36:37Z) - Generative Action Description Prompts for Skeleton-based Action
Recognition [15.38417530693649]
We propose a Generative Action-description Prompts (GAP) approach for skeleton-based action recognition.
We employ a pre-trained large-scale language model as the knowledge engine to automatically generate text descriptions for body parts movements of actions.
Our proposed GAP method achieves noticeable improvements over various baseline models without extra cost at inference.
arXiv Detail & Related papers (2022-08-10T12:55:56Z) - Self-Regulated Learning for Egocentric Video Activity Anticipation [147.9783215348252]
Self-Regulated Learning (SRL) aims to regulate the intermediate representation consecutively to produce representation that emphasizes the novel information in the frame of the current time-stamp.
SRL sharply outperforms existing state-of-the-art in most cases on two egocentric video datasets and two third-person video datasets.
arXiv Detail & Related papers (2021-11-23T03:29:18Z) - Skeleton-Based Mutually Assisted Interacted Object Localization and
Human Action Recognition [111.87412719773889]
We propose a joint learning framework for "interacted object localization" and "human action recognition" based on skeleton data.
Our method achieves the best or competitive performance with the state-of-the-art methods for human action recognition.
arXiv Detail & Related papers (2021-10-28T10:09:34Z) - Skeleton-DML: Deep Metric Learning for Skeleton-Based One-Shot Action
Recognition [0.5161531917413706]
One-shot action recognition allows the recognition of human-performed actions with only a single training example.
This can influence human-robot-interaction positively by enabling the robot to react to previously unseen behaviour.
We propose a novel image-based skeleton representation that performs well in a metric learning setting.
arXiv Detail & Related papers (2020-12-26T22:31:11Z) - Memory-augmented Dense Predictive Coding for Video Representation
Learning [103.69904379356413]
We propose a new architecture and learning framework Memory-augmented Predictive Coding (MemDPC) for the task.
We investigate visual-only self-supervised video representation learning from RGB frames, or from unsupervised optical flow, or both.
In all cases, we demonstrate state-of-the-art or comparable performance over other approaches with orders of magnitude fewer training data.
arXiv Detail & Related papers (2020-08-03T17:57:01Z) - Recognition of Instrument-Tissue Interactions in Endoscopic Videos via
Action Triplets [9.517537672430006]
We tackle the recognition of fine-grained activities, modeled as action triplets instrument, verb, target> representing the tool activity.
We introduce a new laparoscopic dataset, CholecT40, consisting of 40 videos from the public dataset Cholec80 in which all frames have been annotated using 128 triplet classes.
arXiv Detail & Related papers (2020-07-10T14:17:10Z) - Delving into 3D Action Anticipation from Streaming Videos [99.0155538452263]
Action anticipation aims to recognize the action with a partial observation.
We introduce several complementary evaluation metrics and present a basic model based on frame-wise action classification.
We also explore multi-task learning strategies by incorporating auxiliary information from two aspects: the full action representation and the class-agnostic action label.
arXiv Detail & Related papers (2019-06-15T10:30:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.