Augmented Fairness: An Interpretable Model Augmenting Decision-Makers'
Fairness
- URL: http://arxiv.org/abs/2011.08398v1
- Date: Tue, 17 Nov 2020 03:25:44 GMT
- Title: Augmented Fairness: An Interpretable Model Augmenting Decision-Makers'
Fairness
- Authors: Tong Wang and Maytal Saar-Tsechansky
- Abstract summary: We propose a model-agnostic approach for mitigating the prediction bias of a black-box decision-maker.
Our method detects in the feature space where the black-box decision-maker is biased and replaces it with a few short decision rules, acting as a "fair surrogate"
- Score: 10.53972370889201
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a model-agnostic approach for mitigating the prediction bias of a
black-box decision-maker, and in particular, a human decision-maker. Our method
detects in the feature space where the black-box decision-maker is biased and
replaces it with a few short decision rules, acting as a "fair surrogate". The
rule-based surrogate model is trained under two objectives, predictive
performance and fairness. Our model focuses on a setting that is common in
practice but distinct from other literature on fairness. We only have black-box
access to the model, and only a limited set of true labels can be queried under
a budget constraint. We formulate a multi-objective optimization for building a
surrogate model, where we simultaneously optimize for both predictive
performance and bias. To train the model, we propose a novel training algorithm
that combines a nondominated sorting genetic algorithm with active learning. We
test our model on public datasets where we simulate various biased "black-box"
classifiers (decision-makers) and apply our approach for interpretable
augmented fairness.
Related papers
- FairGridSearch: A Framework to Compare Fairness-Enhancing Models [0.0]
This paper focuses on binary classification and proposes FairGridSearch, a novel framework for comparing fairness-enhancing models.
The study applies FairGridSearch to three popular datasets (Adult, COMPAS, and German Credit) and analyzes the impacts of metric selection, base estimator choice, and classification threshold on model fairness.
arXiv Detail & Related papers (2024-01-04T10:29:02Z) - Fast Model Debias with Machine Unlearning [54.32026474971696]
Deep neural networks might behave in a biased manner in many real-world scenarios.
Existing debiasing methods suffer from high costs in bias labeling or model re-training.
We propose a fast model debiasing framework (FMD) which offers an efficient approach to identify, evaluate and remove biases.
arXiv Detail & Related papers (2023-10-19T08:10:57Z) - Evaluating the Fairness of Discriminative Foundation Models in Computer
Vision [51.176061115977774]
We propose a novel taxonomy for bias evaluation of discriminative foundation models, such as Contrastive Language-Pretraining (CLIP)
We then systematically evaluate existing methods for mitigating bias in these models with respect to our taxonomy.
Specifically, we evaluate OpenAI's CLIP and OpenCLIP models for key applications, such as zero-shot classification, image retrieval and image captioning.
arXiv Detail & Related papers (2023-10-18T10:32:39Z) - DualFair: Fair Representation Learning at Both Group and Individual
Levels via Contrastive Self-supervision [73.80009454050858]
This work presents a self-supervised model, called DualFair, that can debias sensitive attributes like gender and race from learned representations.
Our model jointly optimize for two fairness criteria - group fairness and counterfactual fairness.
arXiv Detail & Related papers (2023-03-15T07:13:54Z) - Stochastic Methods for AUC Optimization subject to AUC-based Fairness
Constraints [51.12047280149546]
A direct approach for obtaining a fair predictive model is to train the model through optimizing its prediction performance subject to fairness constraints.
We formulate the training problem of a fairness-aware machine learning model as an AUC optimization problem subject to a class of AUC-based fairness constraints.
We demonstrate the effectiveness of our approach on real-world data under different fairness metrics.
arXiv Detail & Related papers (2022-12-23T22:29:08Z) - A Differentiable Distance Approximation for Fairer Image Classification [31.471917430653626]
We propose a differentiable approximation of the variance of demographics, a metric that can be used to measure the bias, or unfairness, in an AI model.
Our approximation can be optimised alongside the regular training objective which eliminates the need for any extra models during training.
We demonstrate that our approach improves the fairness of AI models in varied task and dataset scenarios.
arXiv Detail & Related papers (2022-10-09T23:02:18Z) - fairmodels: A Flexible Tool For Bias Detection, Visualization, And
Mitigation [3.548416925804316]
This article introduces an R package fairmodels that helps to validate fairness and eliminate bias in classification models.
The implemented set of functions and fairness metrics enables model fairness validation from different perspectives.
The package includes a series of methods for bias mitigation that aim to diminish the discrimination in the model.
arXiv Detail & Related papers (2021-04-01T15:06:13Z) - Design of Dynamic Experiments for Black-Box Model Discrimination [72.2414939419588]
Consider a dynamic model discrimination setting where we wish to chose: (i) what is the best mechanistic, time-varying model and (ii) what are the best model parameter estimates.
For rival mechanistic models where we have access to gradient information, we extend existing methods to incorporate a wider range of problem uncertainty.
We replace these black-box models with Gaussian process surrogate models and thereby extend the model discrimination setting to additionally incorporate rival black-box model.
arXiv Detail & Related papers (2021-02-07T11:34:39Z) - Characterizing Fairness Over the Set of Good Models Under Selective
Labels [69.64662540443162]
We develop a framework for characterizing predictive fairness properties over the set of models that deliver similar overall performance.
We provide tractable algorithms to compute the range of attainable group-level predictive disparities.
We extend our framework to address the empirically relevant challenge of selectively labelled data.
arXiv Detail & Related papers (2021-01-02T02:11:37Z) - Do the Machine Learning Models on a Crowd Sourced Platform Exhibit Bias?
An Empirical Study on Model Fairness [7.673007415383724]
We have created a benchmark of 40 top-rated models from Kaggle used for 5 different tasks.
We have applied 7 mitigation techniques on these models and analyzed the fairness, mitigation results, and impacts on performance.
arXiv Detail & Related papers (2020-05-21T23:35:53Z) - FairALM: Augmented Lagrangian Method for Training Fair Models with
Little Regret [42.66567001275493]
It is now accepted that because of biases in the datasets we present to the models, a fairness-oblivious training will lead to unfair models.
Here, we study mechanisms that impose fairness concurrently while training the model.
arXiv Detail & Related papers (2020-04-03T03:18:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.