Students Success Modeling: Most Important Factors
- URL: http://arxiv.org/abs/2309.13052v1
- Date: Wed, 6 Sep 2023 19:23:10 GMT
- Title: Students Success Modeling: Most Important Factors
- Authors: Sahar Voghoei, James M. Byars, Scott Jackson King, Soheil Shapouri,
Hamed Yaghoobian, Khaled M. Rasheed, Hamid R. Arabnia
- Abstract summary: The model undertakes to identify students likely to graduate, the ones likely to transfer to a different school, and the ones likely to drop out and leave their higher education unfinished.
Our experiments demonstrate that distinguishing between to-be-graduate and at-risk students is reasonably achievable in the earliest stages.
The model remarkably foresees the fate of students who stay in the school for three years.
- Score: 0.47829670123819784
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The importance of retention rate for higher education institutions has
encouraged data analysts to present various methods to predict at-risk
students. The present study, motivated by the same encouragement, proposes a
deep learning model trained with 121 features of diverse categories extracted
or engineered out of the records of 60,822 postsecondary students. The model
undertakes to identify students likely to graduate, the ones likely to transfer
to a different school, and the ones likely to drop out and leave their higher
education unfinished. This study undertakes to adjust its predictive methods
for different stages of curricular progress of students. The temporal aspects
introduced for this purpose are accounted for by incorporating layers of LSTM
in the model. Our experiments demonstrate that distinguishing between
to-be-graduate and at-risk students is reasonably achievable in the earliest
stages, and then it rapidly improves, but the resolution within the latter
category (dropout vs. transfer) depends on data accumulated over time. However,
the model remarkably foresees the fate of students who stay in the school for
three years. The model is also assigned to present the weightiest features in
the procedure of prediction, both on institutional and student levels. A large,
diverse sample size along with the investigation of more than one hundred
extracted or engineered features in our study provide new insights into
variables that affect students success, predict dropouts with reasonable
accuracy, and shed light on the less investigated issue of transfer between
colleges. More importantly, by providing individual-level predictions (as
opposed to school-level predictions) and addressing the outcomes of transfers,
this study improves the use of ML in the prediction of educational outcomes.
Related papers
- Machine Learning Predicts Upper Secondary Education Dropout as Early as the End of Primary School [0.0]
This study expanded the modeling horizon by utilizing a 13-year longitudinal dataset, encompassing data from kindergarten to Grade 9.
Our methodology incorporated a comprehensive range of parameters, including students' academic and cognitive skills, motivation, behavior, well-being, and officially recorded dropout data.
The machine learning models developed in this study demonstrated notable classification ability, achieving a mean area under the curve (AUC) of 0.61 with data up to Grade 6 and an improved AUC of 0.65 with data up to Grade 9.
arXiv Detail & Related papers (2024-03-01T13:18:08Z) - Temporal and Between-Group Variability in College Dropout Prediction [0.0]
This study provides a systematic evaluation of contributing factors and predictive performance of machine learning models.
We find dropout prediction at the end of the second year has a 20% higher AUC than at the time of enrollment in a Random Forest model.
Regarding variability across student groups, college GPA has more predictive value for students from traditionally disadvantaged backgrounds than their peers.
arXiv Detail & Related papers (2024-01-12T10:43:55Z) - Small-scale proxies for large-scale Transformer training instabilities [69.36381318171338]
We seek ways to reproduce and study training stability and instability at smaller scales.
By measuring the relationship between learning rate and loss across scales, we show that these instabilities also appear in small models when training at high learning rates.
We study methods such as warm-up, weight decay, and the $mu$Param to train small models that achieve similar losses across orders of magnitude of learning rate variation.
arXiv Detail & Related papers (2023-09-25T17:48:51Z) - A Predictive Model using Machine Learning Algorithm in Identifying
Students Probability on Passing Semestral Course [0.0]
This study employs classification for data mining techniques, and decision tree for algorithm.
With the utilization of the newly discovered predictive model, the prediction of students probabilities to pass the current courses they take gives 0.7619 accuracy, 0.8333 precision, 0.8823 recall, and 0.8571 f1 score.
arXiv Detail & Related papers (2023-04-12T01:57:08Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Predicting student performance using sequence classification with
time-based windows [1.5836913530330787]
We show that accurate predictive models can be built based on sequential patterns derived from students' behavioral data.
We present a methodology for capturing temporal aspects in behavioral data and analyze its influence on the predictive performance of the models.
The results of our improved sequence classification technique are capable of predicting student performance with high levels of accuracy, reaching 90 percent for course-specific models.
arXiv Detail & Related papers (2022-08-16T13:46:39Z) - LifeLonger: A Benchmark for Continual Disease Classification [59.13735398630546]
We introduce LifeLonger, a benchmark for continual disease classification on the MedMNIST collection.
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch.
Cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
arXiv Detail & Related papers (2022-04-12T12:25:05Z) - Rethinking Importance Weighting for Transfer Learning [71.81262398144946]
Key assumption in supervised learning is that training and test data follow the same probability distribution.
As real-world machine learning tasks are becoming increasingly complex, novel approaches are explored to cope with such challenges.
arXiv Detail & Related papers (2021-12-19T14:35:25Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z) - Predicting MOOCs Dropout Using Only Two Easily Obtainable Features from
the First Week's Activities [56.1344233010643]
Several features are considered to contribute towards learner attrition or lack of interest, which may lead to disengagement or total dropout.
This study aims to predict dropout early-on, from the first week, by comparing several machine-learning approaches.
arXiv Detail & Related papers (2020-08-12T10:44:49Z) - Academic Performance Estimation with Attention-based Graph Convolutional
Networks [17.985752744098267]
Given a student's past data, the task of student's performance prediction is to predict a student's grades in future courses.
Traditional methods for student's performance prediction usually neglect the underlying relationships between multiple courses.
We propose a novel attention-based graph convolutional networks model for student's performance prediction.
arXiv Detail & Related papers (2019-12-26T23:11:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.