Computational Models for Academic Performance Estimation
- URL: http://arxiv.org/abs/2009.02661v1
- Date: Sun, 6 Sep 2020 07:31:37 GMT
- Title: Computational Models for Academic Performance Estimation
- Authors: Vipul Bansal, Himanshu Buckchash, Balasubramanian Raman
- Abstract summary: This paper presents an in-depth analysis of deep learning and machine learning approaches for the formulation of an automated students' performance estimation system.
Our main contributions are (a) a large dataset with fifteen courses (shared publicly for academic research) (b) statistical analysis and ablations on the estimation problem for this dataset.
Unlike previous approaches that rely on feature engineering or logical function deduction, our approach is fully data-driven and thus highly generic with better performance across different prediction tasks.
- Score: 21.31653695065347
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Evaluation of students' performance for the completion of courses has been a
major problem for both students and faculties during the work-from-home period
in this COVID pandemic situation. To this end, this paper presents an in-depth
analysis of deep learning and machine learning approaches for the formulation
of an automated students' performance estimation system that works on partially
available students' academic records. Our main contributions are (a) a large
dataset with fifteen courses (shared publicly for academic research) (b)
statistical analysis and ablations on the estimation problem for this dataset
(c) predictive analysis through deep learning approaches and comparison with
other arts and machine learning algorithms. Unlike previous approaches that
rely on feature engineering or logical function deduction, our approach is
fully data-driven and thus highly generic with better performance across
different prediction tasks.
Related papers
- Research on Education Big Data for Students Academic Performance Analysis based on Machine Learning [8.556825982336807]
In this work, a machine learning model based on Long Short-Term Memory Network (LSTM) was used to conduct an in-depth analysis of educational big data.
The LSTM model efficiently processes time series data, allowing us to capture time-dependent and long-term trends in students' learning activities.
This approach is particularly useful for analyzing student progress, engagement, and other behavioral patterns to support personalized education.
arXiv Detail & Related papers (2024-06-25T01:19:22Z) - Data Shapley in One Training Run [88.59484417202454]
Data Shapley provides a principled framework for attributing data's contribution within machine learning contexts.
Existing approaches require re-training models on different data subsets, which is computationally intensive.
This paper introduces In-Run Data Shapley, which addresses these limitations by offering scalable data attribution for a target model of interest.
arXiv Detail & Related papers (2024-06-16T17:09:24Z) - Learning-Augmented Algorithms with Explicit Predictors [67.02156211760415]
Recent advances in algorithmic design show how to utilize predictions obtained by machine learning models from past and present data.
Prior research in this context was focused on a paradigm where the predictor is pre-trained on past data and then used as a black box.
In this work, we unpack the predictor and integrate the learning problem it gives rise for within the algorithmic challenge.
arXiv Detail & Related papers (2024-03-12T08:40:21Z) - Manipulating Predictions over Discrete Inputs in Machine Teaching [43.914943603238996]
This paper focuses on machine teaching in the discrete domain, specifically on manipulating student models' predictions based on the goals of teachers via changing the training data efficiently.
We formulate this task as a optimization problem and solve it by proposing an iterative searching algorithm.
Our algorithm demonstrates significant numerical merit in the scenarios where a teacher attempts at correcting erroneous predictions to improve the student's models, or maliciously manipulating the model to misclassify some specific samples to the target class aligned with his personal profits.
arXiv Detail & Related papers (2024-01-31T14:23:51Z) - PILOT: A Pre-Trained Model-Based Continual Learning Toolbox [71.63186089279218]
This paper introduces a pre-trained model-based continual learning toolbox known as PILOT.
On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt.
On the other hand, PILOT fits typical class-incremental learning algorithms within the context of pre-trained models to evaluate their effectiveness.
arXiv Detail & Related papers (2023-09-13T17:55:11Z) - Analyzing the Capabilities of Nature-inspired Feature Selection
Algorithms in Predicting Student Performance [0.0]
In this paper, an analysis was conducted to determine the relative performance of a suite of nature-inspired algorithms in the feature-selection portion of ensemble algorithms used to predict student performance.
It was found that leveraging an ensemble approach using nature-inspired algorithms for feature selection and traditional ML algorithms for classification significantly increased predictive accuracy while also reducing feature set size by up to 65 percent.
arXiv Detail & Related papers (2023-08-15T21:18:52Z) - On the Trade-off of Intra-/Inter-class Diversity for Supervised
Pre-training [72.8087629914444]
We study the impact of the trade-off between the intra-class diversity (the number of samples per class) and the inter-class diversity (the number of classes) of a supervised pre-training dataset.
With the size of the pre-training dataset fixed, the best downstream performance comes with a balance on the intra-/inter-class diversity.
arXiv Detail & Related papers (2023-05-20T16:23:50Z) - Representation Learning with Multi-Step Inverse Kinematics: An Efficient
and Optimal Approach to Rich-Observation RL [106.82295532402335]
Existing reinforcement learning algorithms suffer from computational intractability, strong statistical assumptions, and suboptimal sample complexity.
We provide the first computationally efficient algorithm that attains rate-optimal sample complexity with respect to the desired accuracy level.
Our algorithm, MusIK, combines systematic exploration with representation learning based on multi-step inverse kinematics.
arXiv Detail & Related papers (2023-04-12T14:51:47Z) - A Field Guide to Federated Optimization [161.3779046812383]
Federated learning and analytics are a distributed approach for collaboratively learning models (or statistics) from decentralized data.
This paper provides recommendations and guidelines on formulating, designing, evaluating and analyzing federated optimization algorithms.
arXiv Detail & Related papers (2021-07-14T18:09:08Z) - Multi-split Optimized Bagging Ensemble Model Selection for Multi-class
Educational Data Mining [8.26773636337474]
This work analyzes two different undergraduate datasets at two different universities.
It aims to predict the students' performance at two stages of course delivery (20% and 50% respectively)
arXiv Detail & Related papers (2020-06-09T03:22:33Z) - Systematic Ensemble Model Selection Approach for Educational Data Mining [8.26773636337474]
This work explores and analyzing two different datasets at two separate stages of course delivery.
It proposes a systematic approach based on Gini index and p-value to select a suitable ensemble learner from a combination of six potential machine learning algorithms.
Experimental results show that the proposed ensemble models achieve high accuracy and low false positive rate at all stages for both datasets.
arXiv Detail & Related papers (2020-05-13T22:25:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.