Practical Bias Mitigation through Proxy Sensitive Attribute Label
Generation
- URL: http://arxiv.org/abs/2312.15994v1
- Date: Tue, 26 Dec 2023 10:54:15 GMT
- Title: Practical Bias Mitigation through Proxy Sensitive Attribute Label
Generation
- Authors: Bhushan Chaudhary, Anubha Pandey, Deepak Bhatt, Darshika Tiwari
- Abstract summary: We propose a two-stage approach of unsupervised embedding generation followed by clustering to obtain proxy-sensitive labels.
The efficacy of our work relies on the assumption that bias propagates through non-sensitive attributes that are correlated to the sensitive attributes.
Experimental results demonstrate that bias mitigation using existing algorithms such as Fair Mixup and Adversarial Debiasing yields comparable results on derived proxy labels.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Addressing bias in the trained machine learning system often requires access
to sensitive attributes. In practice, these attributes are not available either
due to legal and policy regulations or data unavailability for a given
demographic. Existing bias mitigation algorithms are limited in their
applicability to real-world scenarios as they require access to sensitive
attributes to achieve fairness. In this research work, we aim to address this
bottleneck through our proposed unsupervised proxy-sensitive attribute label
generation technique. Towards this end, we propose a two-stage approach of
unsupervised embedding generation followed by clustering to obtain
proxy-sensitive labels. The efficacy of our work relies on the assumption that
bias propagates through non-sensitive attributes that are correlated to the
sensitive attributes and, when mapped to the high dimensional latent space,
produces clusters of different demographic groups that exist in the data.
Experimental results demonstrate that bias mitigation using existing algorithms
such as Fair Mixup and Adversarial Debiasing yields comparable results on
derived proxy labels when compared against using true sensitive attributes.
Related papers
- Fairness Under Demographic Scarce Regime [7.523105080786704]
We propose a framework to build attribute classifiers that achieve better fairness-accuracy tradeoffs.
We show that enforcing fairness constraints on samples with uncertain sensitive attributes can negatively impact the fairness-accuracy tradeoff.
Our framework can outperform models trained with fairness constraints on the true sensitive attributes in most benchmarks.
arXiv Detail & Related papers (2023-07-24T19:07:34Z) - Towards Assumption-free Bias Mitigation [47.5131072745805]
We propose an assumption-free framework to detect the related attributes automatically by modeling feature interaction for bias mitigation.
Experimental results on four real-world datasets demonstrate that our proposed framework can significantly alleviate unfair prediction behaviors.
arXiv Detail & Related papers (2023-07-09T05:55:25Z) - On the Universal Adversarial Perturbations for Efficient Data-free
Adversarial Detection [55.73320979733527]
We propose a data-agnostic adversarial detection framework, which induces different responses between normal and adversarial samples to UAPs.
Experimental results show that our method achieves competitive detection performance on various text classification tasks.
arXiv Detail & Related papers (2023-06-27T02:54:07Z) - Hyper-parameter Tuning for Fair Classification without Sensitive Attribute Access [12.447577504758485]
We propose a framework to train fair classifiers without access to sensitive attributes on either training or validation data.
We show theoretically and empirically that these proxy labels can be used to maximize fairness under average accuracy constraints.
arXiv Detail & Related papers (2023-02-02T19:45:50Z) - Fairness via Adversarial Attribute Neighbourhood Robust Learning [49.93775302674591]
We propose a principled underlineRobust underlineAdversarial underlineAttribute underlineNeighbourhood (RAAN) loss to debias the classification head.
arXiv Detail & Related papers (2022-10-12T23:39:28Z) - Mitigating Algorithmic Bias with Limited Annotations [65.060639928772]
When sensitive attributes are not disclosed or available, it is needed to manually annotate a small part of the training data to mitigate bias.
We propose Active Penalization Of Discrimination (APOD), an interactive framework to guide the limited annotations towards maximally eliminating the effect of algorithmic bias.
APOD shows comparable performance to fully annotated bias mitigation, which demonstrates that APOD could benefit real-world applications when sensitive information is limited.
arXiv Detail & Related papers (2022-07-20T16:31:19Z) - Semi-FairVAE: Semi-supervised Fair Representation Learning with
Adversarial Variational Autoencoder [92.67156911466397]
We propose a semi-supervised fair representation learning approach based on adversarial variational autoencoder.
We use a bias-aware model to capture inherent bias information on sensitive attribute.
We also use a bias-free model to learn debiased fair representations by using adversarial learning to remove bias information from them.
arXiv Detail & Related papers (2022-04-01T15:57:47Z) - Fairness-Aware Naive Bayes Classifier for Data with Multiple Sensitive
Features [0.0]
We generalise two-naive-Bayes (2NB) into N-naive-Bayes (NNB) to eliminate the simplification of assuming only two sensitive groups in the data.
We investigate its application on data with multiple sensitive features and propose a new constraint and post-processing routine to enforce differential fairness.
arXiv Detail & Related papers (2022-02-23T13:32:21Z) - Fairness via Representation Neutralization [60.90373932844308]
We propose a new mitigation technique, namely, Representation Neutralization for Fairness (RNF)
RNF achieves that fairness by debiasing only the task-specific classification head of DNN models.
Experimental results over several benchmark datasets demonstrate our RNF framework to effectively reduce discrimination of DNN models.
arXiv Detail & Related papers (2021-06-23T22:26:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.