You Can Still Achieve Fairness Without Sensitive Attributes: Exploring
Biases in Non-Sensitive Features
- URL: http://arxiv.org/abs/2104.14537v2
- Date: Sat, 1 May 2021 03:24:58 GMT
- Title: You Can Still Achieve Fairness Without Sensitive Attributes: Exploring
Biases in Non-Sensitive Features
- Authors: Tianxiang Zhao, Enyan Dai, Kai Shu, Suhang Wang
- Abstract summary: We propose a novel framework which simultaneously uses these related features for accurate prediction and regularizing the model to be fair.
Experimental results on real-world datasets demonstrate the effectiveness of the proposed model.
- Score: 29.94644351343916
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Though machine learning models are achieving great success, ex-tensive
studies have exposed their disadvantage of inheriting latent discrimination and
societal bias from the training data, which hinders their adoption on
high-state applications. Thus, many efforts have been taken for developing fair
machine learning models. Most of them require that sensitive attributes are
available during training to learn fair models. However, in many real-world
applications, it is usually infeasible to obtain the sensitive attribute due to
privacy or legal issues, which challenges existing fair classifiers. Though the
sensitive attribute of each data sample is unknown, we observe that there are
usually some non-sensitive features in the training data that are highly
correlated with sensitive attributes, which can be used to alleviate the bias.
Therefore, in this paper, we study a novel problem of exploring features that
are highly correlated with sensitive attributes for learning fair and accurate
classifier without sensitive attributes. We theoretically show that by
minimizing the correlation between these related features and model prediction,
we can learn a fair classifier. Based on this motivation, we propose a novel
framework which simultaneously uses these related features for accurate
prediction and regularizing the model to be fair. In addition, the model can
dynamically adjust the importance weight of each related feature to balance the
contribution of the feature on model classification and fairness. Experimental
results on real-world datasets demonstrate the effectiveness of the proposed
model for learning fair models with high classification accuracy.
Related papers
- Achieve Fairness without Demographics for Dermatological Disease
Diagnosis [17.792332189055223]
We propose a method enabling fair predictions for sensitive attributes during the testing phase without using such information during training.
Inspired by prior work highlighting the impact of feature entanglement on fairness, we enhance the model features by capturing the features related to the sensitive and target attributes.
This ensures that the model can only classify based on the features related to the target attribute without relying on features associated with sensitive attributes.
arXiv Detail & Related papers (2024-01-16T02:49:52Z) - Learning for Counterfactual Fairness from Observational Data [62.43249746968616]
Fairness-aware machine learning aims to eliminate biases of learning models against certain subgroups described by certain protected (sensitive) attributes such as race, gender, and age.
A prerequisite for existing methods to achieve counterfactual fairness is the prior human knowledge of the causal model for the data.
In this work, we address the problem of counterfactually fair prediction from observational data without given causal models by proposing a novel framework CLAIRE.
arXiv Detail & Related papers (2023-07-17T04:08:29Z) - Towards Assumption-free Bias Mitigation [47.5131072745805]
We propose an assumption-free framework to detect the related attributes automatically by modeling feature interaction for bias mitigation.
Experimental results on four real-world datasets demonstrate that our proposed framework can significantly alleviate unfair prediction behaviors.
arXiv Detail & Related papers (2023-07-09T05:55:25Z) - Stubborn Lexical Bias in Data and Models [50.79738900885665]
We use a new statistical method to examine whether spurious patterns in data appear in models trained on the data.
We apply an optimization approach to *reweight* the training data, reducing thousands of spurious correlations.
Surprisingly, though this method can successfully reduce lexical biases in the training data, we still find strong evidence of corresponding bias in the trained models.
arXiv Detail & Related papers (2023-06-03T20:12:27Z) - Fair Classification via Domain Adaptation: A Dual Adversarial Learning
Approach [14.344142985726853]
We study a novel problem of exploring domain adaptation for fair classification.
We propose a new framework that can learn to adapt the sensitive attributes from a source domain for fair classification in the target domain.
arXiv Detail & Related papers (2022-06-08T02:53:18Z) - Semi-FairVAE: Semi-supervised Fair Representation Learning with
Adversarial Variational Autoencoder [92.67156911466397]
We propose a semi-supervised fair representation learning approach based on adversarial variational autoencoder.
We use a bias-aware model to capture inherent bias information on sensitive attribute.
We also use a bias-free model to learn debiased fair representations by using adversarial learning to remove bias information from them.
arXiv Detail & Related papers (2022-04-01T15:57:47Z) - Learning Fair Models without Sensitive Attributes: A Generative Approach [33.196044483534784]
We study a novel problem of learning fair models without sensitive attributes by exploring relevant features.
We propose a probabilistic generative framework to effectively estimate the sensitive attribute from the training data.
Experimental results on real-world datasets show the effectiveness of our framework.
arXiv Detail & Related papers (2022-03-30T15:54:30Z) - FairIF: Boosting Fairness in Deep Learning via Influence Functions with
Validation Set Sensitive Attributes [51.02407217197623]
We propose a two-stage training algorithm named FAIRIF.
It minimizes the loss over the reweighted data set where the sample weights are computed.
We show that FAIRIF yields models with better fairness-utility trade-offs against various types of bias.
arXiv Detail & Related papers (2022-01-15T05:14:48Z) - Fairness in Semi-supervised Learning: Unlabeled Data Help to Reduce
Discrimination [53.3082498402884]
A growing specter in the rise of machine learning is whether the decisions made by machine learning models are fair.
We present a framework of fair semi-supervised learning in the pre-processing phase, including pseudo labeling to predict labels for unlabeled data.
A theoretical decomposition analysis of bias, variance and noise highlights the different sources of discrimination and the impact they have on fairness in semi-supervised learning.
arXiv Detail & Related papers (2020-09-25T05:48:56Z) - Fairness-Aware Learning with Prejudice Free Representations [2.398608007786179]
We propose a novel algorithm that can effectively identify and treat latent discriminating features.
The approach helps to collect discrimination-free features that would improve the model performance.
arXiv Detail & Related papers (2020-02-26T10:06:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.