Who Decides if AI is Fair? The Labels Problem in Algorithmic Auditing
- URL: http://arxiv.org/abs/2111.08723v1
- Date: Tue, 16 Nov 2021 19:00:03 GMT
- Title: Who Decides if AI is Fair? The Labels Problem in Algorithmic Auditing
- Authors: Abhilash Mishra and Yash Gorana
- Abstract summary: We show that fidelity of the ground truth data can lead to spurious differences in performance of ASRs between urban and rural populations.
Our findings highlight how trade-offs between label quality and data annotation costs can complicate algorithmic audits in practice.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Labelled "ground truth" datasets are routinely used to evaluate and audit AI
algorithms applied in high-stakes settings. However, there do not exist widely
accepted benchmarks for the quality of labels in these datasets. We provide
empirical evidence that quality of labels can significantly distort the results
of algorithmic audits in real-world settings. Using data annotators typically
hired by AI firms in India, we show that fidelity of the ground truth data can
lead to spurious differences in performance of ASRs between urban and rural
populations. After a rigorous, albeit expensive, label cleaning process, these
disparities between groups disappear. Our findings highlight how trade-offs
between label quality and data annotation costs can complicate algorithmic
audits in practice. They also emphasize the need for development of
consensus-driven, widely accepted benchmarks for label quality.
Related papers
- How Does Unlabeled Data Provably Help Out-of-Distribution Detection? [63.41681272937562]
Unlabeled in-the-wild data is non-trivial due to the heterogeneity of both in-distribution (ID) and out-of-distribution (OOD) data.
This paper introduces a new learning framework SAL (Separate And Learn) that offers both strong theoretical guarantees and empirical effectiveness.
arXiv Detail & Related papers (2024-02-05T20:36:33Z) - Certification Labels for Trustworthy AI: Insights From an Empirical
Mixed-Method Study [0.0]
This study empirically investigated certification labels as a promising solution.
We demonstrate that labels can significantly increase end-users' trust and willingness to use AI.
However, end-users' preferences for certification labels and their effect on trust and willingness to use AI were more pronounced in high-stake scenarios.
arXiv Detail & Related papers (2023-05-15T09:51:10Z) - Fairness and Bias in Truth Discovery Algorithms: An Experimental
Analysis [7.575734557466221]
Crowd workers may sometimes provide unreliable labels.
Truth discovery (TD) algorithms are applied to determine the consensus labels from conflicting worker responses.
We conduct a systematic study of the bias and fairness of TD algorithms.
arXiv Detail & Related papers (2023-04-25T04:56:35Z) - Beyond Hard Labels: Investigating data label distributions [0.9668407688201357]
We compare the disparity of learning with hard and soft labels for a synthetic and a real-world dataset.
The application of soft labels leads to improved performance and yields a more regular structure of the internal feature space.
arXiv Detail & Related papers (2022-07-13T14:25:30Z) - Debiased Pseudo Labeling in Self-Training [77.83549261035277]
Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets.
To mitigate the requirement for labeled data, self-training is widely used in both academia and industry by pseudo labeling on readily-available unlabeled data.
We propose Debiased, in which the generation and utilization of pseudo labels are decoupled by two independent heads.
arXiv Detail & Related papers (2022-02-15T02:14:33Z) - Confident in the Crowd: Bayesian Inference to Improve Data Labelling in
Crowdsourcing [0.30458514384586394]
We present new techniques to improve the quality of the labels while attempting to reduce the cost.
This paper investigates the use of more sophisticated methods, such as Bayesian inference, to measure the performance of the labellers.
Our methods outperform the standard voting methods in both cost and accuracy while maintaining higher reliability when there is disagreement within the crowd.
arXiv Detail & Related papers (2021-05-28T17:09:45Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - Analysis of label noise in graph-based semi-supervised learning [2.4366811507669124]
In machine learning, one must acquire labels to help supervise a model that will be able to generalize to unseen data.
It is often the case that most of our data is unlabeled.
Semi-supervised learning (SSL) alleviates that by making strong assumptions about the relation between the labels and the input data distribution.
arXiv Detail & Related papers (2020-09-27T22:13:20Z) - Improving Face Recognition by Clustering Unlabeled Faces in the Wild [77.48677160252198]
We propose a novel identity separation method based on extreme value theory.
It greatly reduces the problems caused by overlapping-identity label noise.
Experiments on both controlled and real settings demonstrate our method's consistent improvements.
arXiv Detail & Related papers (2020-07-14T12:26:50Z) - Classify and Generate Reciprocally: Simultaneous Positive-Unlabelled
Learning and Conditional Generation with Extra Data [77.31213472792088]
The scarcity of class-labeled data is a ubiquitous bottleneck in many machine learning problems.
We address this problem by leveraging Positive-Unlabeled(PU) classification and the conditional generation with extra unlabeled data.
We present a novel training framework to jointly target both PU classification and conditional generation when exposed to extra data.
arXiv Detail & Related papers (2020-06-14T08:27:40Z) - Data Augmentation Imbalance For Imbalanced Attribute Classification [60.71438625139922]
We propose a new re-sampling algorithm called: data augmentation imbalance (DAI) to explicitly enhance the ability to discriminate the fewer attributes.
Our DAI algorithm achieves state-of-the-art results, based on pedestrian attribute datasets.
arXiv Detail & Related papers (2020-04-19T20:43:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.