Protected Attributes Tell Us Who, Behavior Tells Us How: A Comparison of
Demographic and Behavioral Oversampling for Fair Student Success Modeling
- URL: http://arxiv.org/abs/2212.10166v1
- Date: Tue, 20 Dec 2022 11:09:11 GMT
- Title: Protected Attributes Tell Us Who, Behavior Tells Us How: A Comparison of
Demographic and Behavioral Oversampling for Fair Student Success Modeling
- Authors: Jade Ma\"i Cock, Muhammad Bilal, Richard Davis, Mirko Marras, Tanja
K\"aser
- Abstract summary: We analyze the fairness of models which use behavioral data to identify at-risk students and suggest two novel pre-processing approaches for bias mitigation.
Based on the concept of intersectionality, the first approach involves intelligent oversampling on combinations of demographic attributes.
The second approach does not require any knowledge of demographic attributes and is based on the assumption that such attributes are a (noisy) proxy for student behavior.
- Score: 6.58879009604603
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Algorithms deployed in education can shape the learning experience and
success of a student. It is therefore important to understand whether and how
such algorithms might create inequalities or amplify existing biases. In this
paper, we analyze the fairness of models which use behavioral data to identify
at-risk students and suggest two novel pre-processing approaches for bias
mitigation. Based on the concept of intersectionality, the first approach
involves intelligent oversampling on combinations of demographic attributes.
The second approach does not require any knowledge of demographic attributes
and is based on the assumption that such attributes are a (noisy) proxy for
student behavior. We hence propose to directly oversample different types of
behaviors identified in a cluster analysis. We evaluate our approaches on data
from (i) an open-ended learning environment and (ii) a flipped classroom
course. Our results show that both approaches can mitigate model bias. Directly
oversampling on behavior is a valuable alternative, when demographic metadata
is not available. Source code and extended results are provided in
https://github.com/epfl-ml4ed/behavioral-oversampling}{https://github.com/epfl-ml4ed/behavioral-oversampling .
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - Leveraging vision-language models for fair facial attribute classification [19.93324644519412]
General-purpose vision-language model (VLM) is a rich knowledge source for common sensitive attributes.
We analyze the correspondence between VLM predicted and human defined sensitive attribute distribution.
Experiments on multiple benchmark facial attribute classification datasets show fairness gains of the model over existing unsupervised baselines.
arXiv Detail & Related papers (2024-03-15T18:37:15Z) - Memory Consistency Guided Divide-and-Conquer Learning for Generalized
Category Discovery [56.172872410834664]
Generalized category discovery (GCD) aims at addressing a more realistic and challenging setting of semi-supervised learning.
We propose a Memory Consistency guided Divide-and-conquer Learning framework (MCDL)
Our method outperforms state-of-the-art models by a large margin on both seen and unseen classes of the generic image recognition.
arXiv Detail & Related papers (2024-01-24T09:39:45Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Learning Bias-Invariant Representation by Cross-Sample Mutual
Information Minimization [77.8735802150511]
We propose a cross-sample adversarial debiasing (CSAD) method to remove the bias information misused by the target task.
The correlation measurement plays a critical role in adversarial debiasing and is conducted by a cross-sample neural mutual information estimator.
We conduct thorough experiments on publicly available datasets to validate the advantages of the proposed method over state-of-the-art approaches.
arXiv Detail & Related papers (2021-08-11T21:17:02Z) - Automatic Fairness Testing of Neural Classifiers through Adversarial
Sampling [8.2868128804393]
We propose a scalable and effective approach for systematically searching for discriminative samples.
Compared with state-of-the-art methods, our approach only employs lightweight procedures like gradient computation and clustering.
The retrained models reduce discrimination by 57.2% and 60.2% respectively on average.
arXiv Detail & Related papers (2021-07-17T03:47:08Z) - Scalable Personalised Item Ranking through Parametric Density Estimation [53.44830012414444]
Learning from implicit feedback is challenging because of the difficult nature of the one-class problem.
Most conventional methods use a pairwise ranking approach and negative samplers to cope with the one-class problem.
We propose a learning-to-rank approach, which achieves convergence speed comparable to the pointwise counterpart.
arXiv Detail & Related papers (2021-05-11T03:38:16Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Discriminative, Generative and Self-Supervised Approaches for
Target-Agnostic Learning [8.666667951130892]
generative and self-supervised learning models are shown to perform well at the task.
Our derived theorem for the pseudo-likelihood theory also shows that they are related for inferring a joint distribution model.
arXiv Detail & Related papers (2020-11-12T15:03:40Z) - A framework for predicting, interpreting, and improving Learning
Outcomes [0.0]
We develop an Embibe Score Quotient model (ESQ) to predict test scores based on observed academic, behavioral and test-taking features of a student.
ESQ can be used to predict the future scoring potential of a student as well as offer personalized learning nudges.
arXiv Detail & Related papers (2020-10-06T11:22:27Z) - Fairness-Aware Learning with Prejudice Free Representations [2.398608007786179]
We propose a novel algorithm that can effectively identify and treat latent discriminating features.
The approach helps to collect discrimination-free features that would improve the model performance.
arXiv Detail & Related papers (2020-02-26T10:06:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.