D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias
- URL: http://arxiv.org/abs/2208.05126v1
- Date: Wed, 10 Aug 2022 03:41:48 GMT
- Title: D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias
- Authors: Bhavya Ghai and Klaus Mueller
- Abstract summary: We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
- Score: 57.87117733071416
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the rise of AI, algorithms have become better at learning underlying
patterns from the training data including ingrained social biases based on
gender, race, etc. Deployment of such algorithms to domains such as hiring,
healthcare, law enforcement, etc. has raised serious concerns about fairness,
accountability, trust and interpretability in machine learning algorithms. To
alleviate this problem, we propose D-BIAS, a visual interactive tool that
embodies human-in-the-loop AI approach for auditing and mitigating social
biases from tabular datasets. It uses a graphical causal model to represent
causal relationships among different features in the dataset and as a medium to
inject domain knowledge. A user can detect the presence of bias against a
group, say females, or a subgroup, say black females, by identifying unfair
causal relationships in the causal network and using an array of fairness
metrics. Thereafter, the user can mitigate bias by acting on the unfair causal
edges. For each interaction, say weakening/deleting a biased causal edge, the
system uses a novel method to simulate a new (debiased) dataset based on the
current causal model. Users can visually assess the impact of their
interactions on different fairness metrics, utility metrics, data distortion,
and the underlying data distribution. Once satisfied, they can download the
debiased dataset and use it for any downstream application for fairer
predictions. We evaluate D-BIAS by conducting experiments on 3 datasets and
also a formal user study. We found that D-BIAS helps reduce bias significantly
compared to the baseline debiasing approach across different fairness metrics
while incurring little data distortion and a small loss in utility. Moreover,
our human-in-the-loop based approach significantly outperforms an automated
approach on trust, interpretability and accountability.
Related papers
- Outlier Detection Bias Busted: Understanding Sources of Algorithmic Bias through Data-centric Factors [28.869581543676947]
unsupervised outlier detection (OD) has numerous applications in finance, security, etc.
This work aims to shed light on the possible sources of unfairness in OD by auditing detection models under different data-centric factors.
We find that the OD algorithms under the study all exhibit fairness pitfalls, although differing in which types of data bias they are more susceptible to.
arXiv Detail & Related papers (2024-08-24T20:35:32Z) - Fast Model Debias with Machine Unlearning [54.32026474971696]
Deep neural networks might behave in a biased manner in many real-world scenarios.
Existing debiasing methods suffer from high costs in bias labeling or model re-training.
We propose a fast model debiasing framework (FMD) which offers an efficient approach to identify, evaluate and remove biases.
arXiv Detail & Related papers (2023-10-19T08:10:57Z) - Learning for Counterfactual Fairness from Observational Data [62.43249746968616]
Fairness-aware machine learning aims to eliminate biases of learning models against certain subgroups described by certain protected (sensitive) attributes such as race, gender, and age.
A prerequisite for existing methods to achieve counterfactual fairness is the prior human knowledge of the causal model for the data.
In this work, we address the problem of counterfactually fair prediction from observational data without given causal models by proposing a novel framework CLAIRE.
arXiv Detail & Related papers (2023-07-17T04:08:29Z) - Understanding Unfairness in Fraud Detection through Model and Data Bias
Interactions [4.159343412286401]
We argue that algorithmic unfairness stems from interactions between models and biases in the data.
We study a set of hypotheses regarding the fairness-accuracy trade-offs that fairness-blind ML algorithms exhibit under different data bias settings.
arXiv Detail & Related papers (2022-07-13T15:18:30Z) - Does Data Repair Lead to Fair Models? Curating Contextually Fair Data To
Reduce Model Bias [10.639605996067534]
Contextual information is a valuable cue for Deep Neural Networks (DNNs) to learn better representations and improve accuracy.
In COCO, many object categories have a much higher co-occurrence with men compared to women, which can bias a DNN's prediction in favor of men.
We introduce a data repair algorithm using the coefficient of variation, which can curate fair and contextually balanced data for a protected class.
arXiv Detail & Related papers (2021-10-20T06:00:03Z) - Balancing out Bias: Achieving Fairness Through Training Reweighting [58.201275105195485]
Bias in natural language processing arises from models learning characteristics of the author such as gender and race.
Existing methods for mitigating and measuring bias do not directly account for correlations between author demographics and linguistic variables.
This paper introduces a very simple but highly effective method for countering bias using instance reweighting.
arXiv Detail & Related papers (2021-09-16T23:40:28Z) - Improving Fairness of AI Systems with Lossless De-biasing [15.039284892391565]
Mitigating bias in AI systems to increase overall fairness has emerged as an important challenge.
We present an information-lossless de-biasing technique that targets the scarcity of data in the disadvantaged group.
arXiv Detail & Related papers (2021-05-10T17:38:38Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Balancing Biases and Preserving Privacy on Balanced Faces in the Wild [50.915684171879036]
There are demographic biases present in current facial recognition (FR) models.
We introduce our Balanced Faces in the Wild dataset to measure these biases across different ethnic and gender subgroups.
We find that relying on a single score threshold to differentiate between genuine and imposters sample pairs leads to suboptimal results.
We propose a novel domain adaptation learning scheme that uses facial features extracted from state-of-the-art neural networks.
arXiv Detail & Related papers (2021-03-16T15:05:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.