Ex-Ante Assessment of Discrimination in Dataset
- URL: http://arxiv.org/abs/2208.07918v2
- Date: Thu, 18 Aug 2022 14:35:41 GMT
- Title: Ex-Ante Assessment of Discrimination in Dataset
- Authors: Jonathan Vasquez, Xavier Gitiaux and Huzefa Rangwala
- Abstract summary: Data owners face increasing liability for how the use of their data could harm under-priviliged communities.
We propose FORESEE, a FORESt of decision trEEs algorithm, which generates a score that captures how likely an individual's response varies with sensitive attributes.
- Score: 20.574371560492494
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data owners face increasing liability for how the use of their data could
harm under-priviliged communities. Stakeholders would like to identify the
characteristics of data that lead to algorithms being biased against any
particular demographic groups, for example, defined by their race, gender, age,
and/or religion. Specifically, we are interested in identifying subsets of the
feature space where the ground truth response function from features to
observed outcomes differs across demographic groups. To this end, we propose
FORESEE, a FORESt of decision trEEs algorithm, which generates a score that
captures how likely an individual's response varies with sensitive attributes.
Empirically, we find that our approach allows us to identify the individuals
who are most likely to be misclassified by several classifiers, including
Random Forest, Logistic Regression, Support Vector Machine, and k-Nearest
Neighbors. The advantage of our approach is that it allows stakeholders to
characterize risky samples that may contribute to discrimination, as well as,
use the FORESEE to estimate the risk of upcoming samples.
Related papers
- Learning under Selective Labels with Data from Heterogeneous
Decision-makers: An Instrumental Variable Approach [7.629248625993988]
We study the problem of learning with selectively labeled data, which arises when outcomes are only partially labeled due to historical decision-making.
We propose a weighted learning approach that learns prediction rules robust to the label selection bias in both identification settings.
arXiv Detail & Related papers (2023-06-13T06:34:44Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Bounding Counterfactuals under Selection Bias [60.55840896782637]
We propose a first algorithm to address both identifiable and unidentifiable queries.
We prove that, in spite of the missingness induced by the selection bias, the likelihood of the available data is unimodal.
arXiv Detail & Related papers (2022-07-26T10:33:10Z) - Data Representativeness in Accessibility Datasets: A Meta-Analysis [7.6597163467929805]
We review datasets sourced by people with disabilities and older adults.
We find that accessibility datasets represent diverse ages, but have gender and race representation gaps.
We hope our effort expands the space of possibility for greater inclusion of marginalized communities in AI-infused systems.
arXiv Detail & Related papers (2022-07-16T23:32:19Z) - Demographic-Reliant Algorithmic Fairness: Characterizing the Risks of
Demographic Data Collection in the Pursuit of Fairness [0.0]
We consider calls to collect more data on demographics to enable algorithmic fairness.
We show how these techniques largely ignore broader questions of data governance and systemic oppression.
arXiv Detail & Related papers (2022-04-18T04:50:09Z) - Measuring Fairness Under Unawareness of Sensitive Attributes: A
Quantification-Based Approach [131.20444904674494]
We tackle the problem of measuring group fairness under unawareness of sensitive attributes.
We show that quantification approaches are particularly suited to tackle the fairness-under-unawareness problem.
arXiv Detail & Related papers (2021-09-17T13:45:46Z) - Information Theoretic Measures for Fairness-aware Feature Selection [27.06618125828978]
We develop a framework for fairness-aware feature selection, based on information theoretic measures for the accuracy and discriminatory impacts of features.
Specifically, our goal is to design a fairness utility score for each feature which quantifies how this feature influences accurate as well as nondiscriminatory decisions.
arXiv Detail & Related papers (2021-06-01T20:11:54Z) - Balancing Biases and Preserving Privacy on Balanced Faces in the Wild [50.915684171879036]
There are demographic biases present in current facial recognition (FR) models.
We introduce our Balanced Faces in the Wild dataset to measure these biases across different ethnic and gender subgroups.
We find that relying on a single score threshold to differentiate between genuine and imposters sample pairs leads to suboptimal results.
We propose a novel domain adaptation learning scheme that uses facial features extracted from state-of-the-art neural networks.
arXiv Detail & Related papers (2021-03-16T15:05:49Z) - Differentially Private and Fair Deep Learning: A Lagrangian Dual
Approach [54.32266555843765]
This paper studies a model that protects the privacy of the individuals sensitive information while also allowing it to learn non-discriminatory predictors.
The method relies on the notion of differential privacy and the use of Lagrangian duality to design neural networks that can accommodate fairness constraints.
arXiv Detail & Related papers (2020-09-26T10:50:33Z) - Mitigating Face Recognition Bias via Group Adaptive Classifier [53.15616844833305]
This work aims to learn a fair face representation, where faces of every group could be more equally represented.
Our work is able to mitigate face recognition bias across demographic groups while maintaining the competitive accuracy.
arXiv Detail & Related papers (2020-06-13T06:43:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.