xFAIR: Better Fairness via Model-based Rebalancing of Protected
Attributes
- URL: http://arxiv.org/abs/2110.01109v1
- Date: Sun, 3 Oct 2021 22:10:14 GMT
- Title: xFAIR: Better Fairness via Model-based Rebalancing of Protected
Attributes
- Authors: Kewen Peng, Joymallya Chakraborty, Tim Menzies
- Abstract summary: Machine learning software can generate models that inappropriately discriminate against specific protected social groups.
We propose xFAIR, a model-based extrapolation method, that is capable of both mitigating bias and explaining the cause.
- Score: 15.525314212209564
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning software can generate models that inappropriately
discriminate against specific protected social groups (e.g., groups based on
gender, ethnicity, etc). Motivated by those results, software engineering
researchers have proposed many methods for mitigating those discriminatory
effects. While those methods are effective in mitigating bias, few of them can
provide explanations on what is the cause of bias. Here we propose xFAIR, a
model-based extrapolation method, that is capable of both mitigating bias and
explaining the cause. In our xFAIR approach, protected attributes are
represented by models learned from the other independent variables (and these
models offer extrapolations over the space between existing examples). We then
use the extrapolation models to relabel protected attributes, which aims to
offset the biased predictions of the classification model via rebalancing the
distribution of protected attributes. The experiments of this paper show that,
without compromising(original) model performance,xFAIRcan achieve significantly
better group and individual fairness (as measured in different metrics)than
benchmark methods. Moreover, when compared to another instance-based
rebalancing method, our model-based approach shows faster runtime and thus
better scalability
Related papers
- Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Distributionally Generative Augmentation for Fair Facial Attribute Classification [69.97710556164698]
Facial Attribute Classification (FAC) holds substantial promise in widespread applications.
FAC models trained by traditional methodologies can be unfair by exhibiting accuracy inconsistencies across varied data subpopulations.
This work proposes a novel, generation-based two-stage framework to train a fair FAC model on biased data without additional annotation.
arXiv Detail & Related papers (2024-03-11T10:50:53Z) - When Fairness Meets Privacy: Exploring Privacy Threats in Fair Binary Classifiers via Membership Inference Attacks [17.243744418309593]
We propose an efficient MIA method against fairness-enhanced models based on fairness discrepancy results.
We also explore potential strategies for mitigating privacy leakages.
arXiv Detail & Related papers (2023-11-07T10:28:17Z) - Non-Invasive Fairness in Learning through the Lens of Data Drift [88.37640805363317]
We show how to improve the fairness of Machine Learning models without altering the data or the learning algorithm.
We use a simple but key insight: the divergence of trends between different populations, and, consecutively, between a learned model and minority populations, is analogous to data drift.
We explore two strategies (model-splitting and reweighing) to resolve this drift, aiming to improve the overall conformance of models to the underlying data.
arXiv Detail & Related papers (2023-03-30T17:30:42Z) - Investigating Ensemble Methods for Model Robustness Improvement of Text
Classifiers [66.36045164286854]
We analyze a set of existing bias features and demonstrate there is no single model that works best for all the cases.
By choosing an appropriate bias model, we can obtain a better robustness result than baselines with a more sophisticated model design.
arXiv Detail & Related papers (2022-10-28T17:52:10Z) - Simultaneous Improvement of ML Model Fairness and Performance by
Identifying Bias in Data [1.76179873429447]
We propose a data preprocessing technique that can detect instances ascribing a specific kind of bias that should be removed from the dataset before training.
In particular, we claim that in the problem settings where instances exist with similar feature but different labels caused by variation in protected attributes, an inherent bias gets induced in the dataset.
arXiv Detail & Related papers (2022-10-24T13:04:07Z) - Cross-model Fairness: Empirical Study of Fairness and Ethics Under Model Multiplicity [10.144058870887061]
We argue that individuals can be harmed when one predictor is chosen ad hoc from a group of equally well performing models.
Our findings suggest that such unfairness can be readily found in real life and it may be difficult to mitigate by technical means alone.
arXiv Detail & Related papers (2022-03-14T14:33:39Z) - fairmodels: A Flexible Tool For Bias Detection, Visualization, And
Mitigation [3.548416925804316]
This article introduces an R package fairmodels that helps to validate fairness and eliminate bias in classification models.
The implemented set of functions and fairness metrics enables model fairness validation from different perspectives.
The package includes a series of methods for bias mitigation that aim to diminish the discrimination in the model.
arXiv Detail & Related papers (2021-04-01T15:06:13Z) - Characterizing Fairness Over the Set of Good Models Under Selective
Labels [69.64662540443162]
We develop a framework for characterizing predictive fairness properties over the set of models that deliver similar overall performance.
We provide tractable algorithms to compute the range of attainable group-level predictive disparities.
We extend our framework to address the empirically relevant challenge of selectively labelled data.
arXiv Detail & Related papers (2021-01-02T02:11:37Z) - Estimating Generalization under Distribution Shifts via Domain-Invariant
Representations [75.74928159249225]
We use a set of domain-invariant predictors as a proxy for the unknown, true target labels.
The error of the resulting risk estimate depends on the target risk of the proxy model.
arXiv Detail & Related papers (2020-07-06T17:21:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.