Who will dropout from university? Academic risk prediction based on
interpretable machine learning
- URL: http://arxiv.org/abs/2112.01079v1
- Date: Thu, 2 Dec 2021 09:43:31 GMT
- Title: Who will dropout from university? Academic risk prediction based on
interpretable machine learning
- Authors: Shudong Yang (1) ((1) Dalian University of Technology)
- Abstract summary: It predicts academic risk based on the LightGBM model and the interpretable machine learning method of Shapley value.
From the local perspective, the factors affecting academic risk vary from person to person.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In the institutional research mode, in order to explore which characteristics
are the best indicators for predicting academic risk from the student behavior
data sets that have high-dimensional, unbalanced classified small sample, it
transforms the academic risk prediction of college students into a binary
classification task. It predicts academic risk based on the LightGBM model and
the interpretable machine learning method of Shapley value. The simulation
results show that from the global perspective of the prediction model,
characteristics such as the quality of academic partners, the seating position
in classroom, the dormitory study atmosphere, the English scores of the college
entrance examination, the quantity of academic partners, the addiction level of
video games, the mobility of academic partners, and the degree of truancy are
the best 8 predictors for academic risk. It is contrary to intuition that
characteristics such as living in campus or not, work-study, lipstick
addiction, student leader or not, lover amount, and smoking have little
correlation with university academic risk in this experiment. From the local
perspective of the sample, the factors affecting academic risk vary from person
to person. It can perform personalized interpretable analysis through Shapley
values, which cannot be done by traditional mathematical statistical prediction
models. The academic contributions of this research are mainly in two aspects:
First, the learning interaction networks is proposed for the first time, so
that social behavior can be used to compensate for the one-sided individual
behavior and improve the performance of academic risk prediction. Second, the
introduction of Shapley value calculation makes machine learning that lacks a
clear reasoning process visualized, and provides intuitive decision support for
education managers.
Related papers
- Beyond Expectations: Learning with Stochastic Dominance Made Practical [88.06211893690964]
dominance models risk-averse preferences for decision making with uncertain outcomes.
Despite theoretically appealing, the application of dominance in machine learning has been scarce.
We first generalize the dominance concept to enable feasible comparisons between any arbitrary pair of random variables.
We then develop a simple and efficient approach for finding the optimal solution in terms of dominance.
arXiv Detail & Related papers (2024-02-05T03:21:23Z) - Learning for Counterfactual Fairness from Observational Data [62.43249746968616]
Fairness-aware machine learning aims to eliminate biases of learning models against certain subgroups described by certain protected (sensitive) attributes such as race, gender, and age.
A prerequisite for existing methods to achieve counterfactual fairness is the prior human knowledge of the causal model for the data.
In this work, we address the problem of counterfactually fair prediction from observational data without given causal models by proposing a novel framework CLAIRE.
arXiv Detail & Related papers (2023-07-17T04:08:29Z) - Empirical Estimates on Hand Manipulation are Recoverable: A Step Towards
Individualized and Explainable Robotic Support in Everyday Activities [80.37857025201036]
Key challenge for robotic systems is to figure out the behavior of another agent.
Processing correct inferences is especially challenging when (confounding) factors are not controlled experimentally.
We propose equipping robots with the necessary tools to conduct observational studies on people.
arXiv Detail & Related papers (2022-01-27T22:15:56Z) - Balanced Q-learning: Combining the Influence of Optimistic and
Pessimistic Targets [74.04426767769785]
We show that specific types of biases may be preferable, depending on the scenario.
We design a novel reinforcement learning algorithm, Balanced Q-learning, in which the target is modified to be a convex combination of a pessimistic and an optimistic term.
arXiv Detail & Related papers (2021-11-03T07:30:19Z) - Statistical discrimination in learning agents [64.78141757063142]
Statistical discrimination emerges in agent policies as a function of both the bias in the training population and of agent architecture.
We show that less discrimination emerges with agents that use recurrent neural networks, and when their training environment has less bias.
arXiv Detail & Related papers (2021-10-21T18:28:57Z) - Tri-Branch Convolutional Neural Networks for Top-$k$ Focused Academic
Performance Prediction [28.383922154797315]
Academic performance prediction aims to leverage student-related information to predict their future academic outcomes.
In this paper, we analyze students' daily behavior trajectories, which can be comprehensively tracked with campus smartcard records.
We propose a novel Tri-Branch CNN architecture, which is equipped with row-wise, column-wise, and depth-wise convolution and attention operations.
arXiv Detail & Related papers (2021-07-22T02:35:36Z) - Should College Dropout Prediction Models Include Protected Attributes? [0.4125187280299248]
We build machine learning models to predict student dropout after one academic year.
We compare the overall performance and fairness of model predictions with or without four protected attributes.
We find that including protected attributes does not impact the overall prediction performance and it only marginally improves algorithmic fairness of predictions.
arXiv Detail & Related papers (2021-03-28T22:47:30Z) - Predicting Early Dropout: Calibration and Algorithmic Fairness
Considerations [2.7048165023994057]
We develop a machine learning method to predict the risks of university dropout and underperformance.
We analyze if this method leads to discriminatory outcomes for some sensitive groups in terms of prediction accuracy (AUC) and error rates (Generalized False Positive Rate, GFPR, or Generalized False Negative Rate, GFNR)
arXiv Detail & Related papers (2021-03-16T13:42:16Z) - Double Robust Representation Learning for Counterfactual Prediction [68.78210173955001]
We propose a novel scalable method to learn double-robust representations for counterfactual predictions.
We make robust and efficient counterfactual predictions for both individual and average treatment effects.
The algorithm shows competitive performance with the state-of-the-art on real world and synthetic data.
arXiv Detail & Related papers (2020-10-15T16:39:26Z) - A framework for predicting, interpreting, and improving Learning
Outcomes [0.0]
We develop an Embibe Score Quotient model (ESQ) to predict test scores based on observed academic, behavioral and test-taking features of a student.
ESQ can be used to predict the future scoring potential of a student as well as offer personalized learning nudges.
arXiv Detail & Related papers (2020-10-06T11:22:27Z) - Academic Performance Estimation with Attention-based Graph Convolutional
Networks [17.985752744098267]
Given a student's past data, the task of student's performance prediction is to predict a student's grades in future courses.
Traditional methods for student's performance prediction usually neglect the underlying relationships between multiple courses.
We propose a novel attention-based graph convolutional networks model for student's performance prediction.
arXiv Detail & Related papers (2019-12-26T23:11:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.