Demographic-Reliant Algorithmic Fairness: Characterizing the Risks of
Demographic Data Collection in the Pursuit of Fairness
- URL: http://arxiv.org/abs/2205.01038v2
- Date: Wed, 4 May 2022 17:25:56 GMT
- Title: Demographic-Reliant Algorithmic Fairness: Characterizing the Risks of
Demographic Data Collection in the Pursuit of Fairness
- Authors: McKane Andrus and Sarah Villeneuve
- Abstract summary: We consider calls to collect more data on demographics to enable algorithmic fairness.
We show how these techniques largely ignore broader questions of data governance and systemic oppression.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most proposed algorithmic fairness techniques require access to data on a
"sensitive attribute" or "protected category" (such as race, ethnicity, gender,
or sexuality) in order to make performance comparisons and standardizations
across groups, however this data is largely unavailable in practice, hindering
the widespread adoption of algorithmic fairness. Through this paper, we
consider calls to collect more data on demographics to enable algorithmic
fairness and challenge the notion that discrimination can be overcome with
smart enough technical methods and sufficient data alone. We show how these
techniques largely ignore broader questions of data governance and systemic
oppression when categorizing individuals for the purpose of fairer algorithmic
processing. In this work, we explore under what conditions demographic data
should be collected and used to enable algorithmic fairness methods by
characterizing a range of social risks to individuals and communities. For the
risks to individuals we consider the unique privacy risks associated with the
sharing of sensitive attributes likely to be the target of fairness analysis,
the possible harms stemming from miscategorizing and misrepresenting
individuals in the data collection process, and the use of sensitive data
beyond data subjects' expectations. Looking more broadly, the risks to entire
groups and communities include the expansion of surveillance infrastructure in
the name of fairness, misrepresenting and mischaracterizing what it means to be
part of a demographic group or to hold a certain identity, and ceding the
ability to define for themselves what constitutes biased or unfair treatment.
We argue that, by confronting these questions before and during the collection
of demographic data, algorithmic fairness methods are more likely to actually
mitigate harmful treatment disparities without reinforcing systems of
oppression.
Related papers
- The Impact of Differential Feature Under-reporting on Algorithmic Fairness [86.275300739926]
We present an analytically tractable model of differential feature under-reporting.
We then use to characterize the impact of this kind of data bias on algorithmic fairness.
Our results show that, in real world data settings, under-reporting typically leads to increasing disparities.
arXiv Detail & Related papers (2024-01-16T19:16:22Z) - Human-Centric Multimodal Machine Learning: Recent Advances and Testbed
on AI-based Recruitment [66.91538273487379]
There is a certain consensus about the need to develop AI applications with a Human-Centric approach.
Human-Centric Machine Learning needs to be developed based on four main requirements: (i) utility and social good; (ii) privacy and data ownership; (iii) transparency and accountability; and (iv) fairness in AI-driven decision-making processes.
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
arXiv Detail & Related papers (2023-02-13T16:44:44Z) - Ex-Ante Assessment of Discrimination in Dataset [20.574371560492494]
Data owners face increasing liability for how the use of their data could harm under-priviliged communities.
We propose FORESEE, a FORESt of decision trEEs algorithm, which generates a score that captures how likely an individual's response varies with sensitive attributes.
arXiv Detail & Related papers (2022-08-16T19:28:22Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Data Representativeness in Accessibility Datasets: A Meta-Analysis [7.6597163467929805]
We review datasets sourced by people with disabilities and older adults.
We find that accessibility datasets represent diverse ages, but have gender and race representation gaps.
We hope our effort expands the space of possibility for greater inclusion of marginalized communities in AI-infused systems.
arXiv Detail & Related papers (2022-07-16T23:32:19Z) - Fairness-Aware Naive Bayes Classifier for Data with Multiple Sensitive
Features [0.0]
We generalise two-naive-Bayes (2NB) into N-naive-Bayes (NNB) to eliminate the simplification of assuming only two sensitive groups in the data.
We investigate its application on data with multiple sensitive features and propose a new constraint and post-processing routine to enforce differential fairness.
arXiv Detail & Related papers (2022-02-23T13:32:21Z) - Measuring Fairness Under Unawareness of Sensitive Attributes: A
Quantification-Based Approach [131.20444904674494]
We tackle the problem of measuring group fairness under unawareness of sensitive attributes.
We show that quantification approaches are particularly suited to tackle the fairness-under-unawareness problem.
arXiv Detail & Related papers (2021-09-17T13:45:46Z) - MultiFair: Multi-Group Fairness in Machine Learning [52.24956510371455]
We study multi-group fairness in machine learning (MultiFair)
We propose a generic end-to-end algorithmic framework to solve it.
Our proposed framework is generalizable to many different settings.
arXiv Detail & Related papers (2021-05-24T02:30:22Z) - "What We Can't Measure, We Can't Understand": Challenges to Demographic
Data Procurement in the Pursuit of Fairness [0.0]
algorithmic fairness practitioners often do not have access to demographic data they feel they need to detect bias in practice.
We investigated this dilemma through semi-structured interviews with 38 practitioners and professionals either working in or adjacent to algorithmic fairness.
Participants painted a complex picture of what demographic data availability and use look like on the ground.
arXiv Detail & Related papers (2020-10-30T21:06:41Z) - Enhancing Facial Data Diversity with Style-based Face Aging [59.984134070735934]
In particular, face datasets are typically biased in terms of attributes such as gender, age, and race.
We propose a novel, generative style-based architecture for data augmentation that captures fine-grained aging patterns.
We show that the proposed method outperforms state-of-the-art algorithms for age transfer.
arXiv Detail & Related papers (2020-06-06T21:53:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.