Formulating Module Assessment for Improved Academic Performance
Predictability in Higher Education
- URL: http://arxiv.org/abs/2008.13255v1
- Date: Sun, 30 Aug 2020 19:42:31 GMT
- Title: Formulating Module Assessment for Improved Academic Performance
Predictability in Higher Education
- Authors: Mohammed Alsuwaiket, Anas H. Blasi, Ra'Fat Al-Msie'deen
- Abstract summary: This paper proposes a different data preparation process through investigating more than 230000 student records.
The effect of CAR on prediction process using the random forest classification technique has been investigated.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Various studies have shown that students tend to get higher marks when
assessed through coursework based assessment methods which include either
modules that are fully assessed through coursework or a mixture of coursework
and examinations than assessed by examination alone. There are a large number
of educational data mining studies that preprocess data through conventional
data mining processes including data preparation process, but they are using
transcript data as they stand without looking at examination and coursework
results weighting which could affect prediction accuracy. This paper proposes a
different data preparation process through investigating more than 230000
student records in order to prepare students marks based on the assessment
methods of enrolled modules. The data have been processed through different
stages in order to extract a categorical factor through which students module
marks are refined during the data preparation process. The results of this work
show that students final marks should not be isolated from the nature of the
enrolled modules assessment methods. They must rather be investigated
thoroughly and considered during EDMs data preprocessing phases. More
generally, it is concluded that educational data should not be prepared in the
same way as other data types due to differences as data sources, applications,
and types of errors in them. Therefore, an attribute, coursework assessment
ratio, is proposed to be used in order to take the different modules assessment
methods into account while preparing student transcript data. The effect of CAR
on prediction process using the random forest classification technique has been
investigated. It is shown that considering CAR as an attribute increases the
accuracy of predicting students second year averages based on their first year
results.
Related papers
- A Fair Post-Processing Method based on the MADD Metric for Predictive Student Models [1.055551340663609]
A new metric has been developed to evaluate algorithmic fairness in predictive student models.
In this paper, we develop a post-processing method that aims at improving the fairness while preserving the accuracy of relevant predictive models' results.
We experiment with our approach on the task of predicting student success in an online course, using both simulated and real-world educational data.
arXiv Detail & Related papers (2024-07-07T14:53:41Z) - Extracting Training Data from Unconditional Diffusion Models [76.85077961718875]
diffusion probabilistic models (DPMs) are being employed as mainstream models for generative artificial intelligence (AI)
We aim to establish a theoretical understanding of memorization in DPMs with 1) a memorization metric for theoretical analysis, 2) an analysis of conditional memorization with informative and random labels, and 3) two better evaluation metrics for measuring memorization.
Based on the theoretical analysis, we propose a novel data extraction method called textbfSurrogate condItional Data Extraction (SIDE) that leverages a trained on generated data as a surrogate condition to extract training data directly from unconditional diffusion models.
arXiv Detail & Related papers (2024-06-18T16:20:12Z) - Distilled Datamodel with Reverse Gradient Matching [74.75248610868685]
We introduce an efficient framework for assessing data impact, comprising offline training and online evaluation stages.
Our proposed method achieves comparable model behavior evaluation while significantly speeding up the process compared to the direct retraining method.
arXiv Detail & Related papers (2024-04-22T09:16:14Z) - Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals [91.59906995214209]
We propose a new evaluation method, Counterfactual Attentiveness Test (CAT)
CAT uses counterfactuals by replacing part of the input with its counterpart from a different example, expecting an attentive model to change its prediction.
We show that GPT3 becomes less attentive with an increased number of demonstrations, while its accuracy on the test data improves.
arXiv Detail & Related papers (2023-11-16T06:27:35Z) - A Predictive Model using Machine Learning Algorithm in Identifying
Students Probability on Passing Semestral Course [0.0]
This study employs classification for data mining techniques, and decision tree for algorithm.
With the utilization of the newly discovered predictive model, the prediction of students probabilities to pass the current courses they take gives 0.7619 accuracy, 0.8333 precision, 0.8823 recall, and 0.8571 f1 score.
arXiv Detail & Related papers (2023-04-12T01:57:08Z) - A Comprehensive Survey on Test-Time Adaptation under Distribution Shifts [143.14128737978342]
Test-time adaptation, an emerging paradigm, has the potential to adapt a pre-trained model to unlabeled data during testing, before making predictions.
Recent progress in this paradigm highlights the significant benefits of utilizing unlabeled data for training self-adapted models prior to inference.
arXiv Detail & Related papers (2023-03-27T16:32:21Z) - Revisiting Long-tailed Image Classification: Survey and Benchmarks with
New Evaluation Metrics [88.39382177059747]
A corpus of metrics is designed for measuring the accuracy, robustness, and bounds of algorithms for learning with long-tailed distribution.
Based on our benchmarks, we re-evaluate the performance of existing methods on CIFAR10 and CIFAR100 datasets.
arXiv Detail & Related papers (2023-02-03T02:40:54Z) - Process-BERT: A Framework for Representation Learning on Educational
Process Data [68.8204255655161]
We propose a framework for learning representations of educational process data.
Our framework consists of a pre-training step that uses BERT-type objectives to learn representations from sequential process data.
We apply our framework to the 2019 nation's report card data mining competition dataset.
arXiv Detail & Related papers (2022-04-28T16:07:28Z) - Data-Centric Machine Learning in the Legal Domain [0.2624902795082451]
This paper explores how changes in a data set influence the measured performance of a model.
Using three publicly available data sets from the legal domain, we investigate how changes to their size, the train/test splits, and the human labelling accuracy impact the performance.
The observed effects are surprisingly pronounced, especially when the per-class performance is considered.
arXiv Detail & Related papers (2022-01-17T23:05:14Z) - Measuring the Credibility of Student Attendance Data in Higher Education
for Data Mining [0.0]
Student attendance in higher education has always been dealt with in a classical way.
This study tries to formulate the extracted knowledge in a way that guarantees achieving accurate and credible results.
arXiv Detail & Related papers (2020-09-01T20:21:46Z) - Refining Student Marks based on Enrolled Modules Assessment Methods
using Data Mining Techniques [0.0]
We propose a different data preparation process by investigating more than 230000 student records for the preparation of scores.
The effect of Module Assessment Index on the prediction process using Random Forest and Naive Bayes classification techniques were investigated.
arXiv Detail & Related papers (2020-08-30T19:47:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.