Local Law 144: A Critical Analysis of Regression Metrics
- URL: http://arxiv.org/abs/2302.04119v1
- Date: Wed, 8 Feb 2023 15:21:14 GMT
- Title: Local Law 144: A Critical Analysis of Regression Metrics
- Authors: Giulio Filippi, Sara Zannone, Airlie Hilliard, Adriano Koshiyama
- Abstract summary: In November 2021, the New York City Council passed a legislation that mandates bias audits of automated employment decision tools.
From 15th April 2023, companies that use automated tools for hiring or promoting employees are required to have these systems audited.
We argue that both metrics fail to capture distributional differences over the whole domain, and therefore cannot reliably detect bias.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The use of automated decision tools in recruitment has received an increasing
amount of attention. In November 2021, the New York City Council passed a
legislation (Local Law 144) that mandates bias audits of Automated Employment
Decision Tools. From 15th April 2023, companies that use automated tools for
hiring or promoting employees are required to have these systems audited by an
independent entity. Auditors are asked to compute bias metrics that compare
outcomes for different groups, based on sex/gender and race/ethnicity
categories at a minimum. Local Law 144 proposes novel bias metrics for
regression tasks (scenarios where the automated system scores candidates with a
continuous range of values). A previous version of the legislation proposed a
bias metric that compared the mean scores of different groups. The new revised
bias metric compares the proportion of candidates in each group that falls
above the median. In this paper, we argue that both metrics fail to capture
distributional differences over the whole domain, and therefore cannot reliably
detect bias. We first introduce two metrics, as possible alternatives to the
legislation metrics. We then compare these metrics over a range of theoretical
examples, for which the legislation proposed metrics seem to underestimate
bias. Finally, we study real data and show that the legislation metrics can
similarly fail in a real-world recruitment application.
Related papers
- What we learned while automating bias detection in AI hiring systems for compliance with NYC Local Law 144 [0.0]
New York City's Local Law 144 requires employers to conduct independent bias audits for any automated employment decision tools (AEDTs) used in hiring processes.
The law outlines a minimum set of bias tests that AI developers and implementers must perform to ensure compliance.
We have collected and analyzed audits conducted under this law, identified best practices, and developed a software tool to streamline employer compliance.
arXiv Detail & Related papers (2024-12-13T14:14:26Z) - Comprehensive Equity Index (CEI): Definition and Application to Bias Evaluation in Biometrics [47.762333925222926]
We present a novel metric to quantify biased behaviors of machine learning models.
We focus on and apply it to the operational evaluation of face recognition systems.
arXiv Detail & Related papers (2024-09-03T14:19:38Z) - Measuring and Addressing Indexical Bias in Information Retrieval [69.7897730778898]
PAIR framework supports automatic bias audits for ranked documents or entire IR systems.
After introducing DUO, we run an extensive evaluation of 8 IR systems on a new corpus of 32k synthetic and 4.7k natural documents.
A human behavioral study validates our approach, showing that our bias metric can help predict when and how indexical bias will shift a reader's opinion.
arXiv Detail & Related papers (2024-06-06T17:42:37Z) - A Principled Approach for a New Bias Measure [7.352247786388098]
We propose the definition of Uniform Bias (UB), the first bias measure with a clear and simple interpretation in the full range of bias values.
Our results are experimentally validated using nine publicly available datasets and theoretically analyzed, which provide novel insights about the problem.
Based on our approach, we also design a bias mitigation model that might be useful to policymakers.
arXiv Detail & Related papers (2024-05-20T18:14:33Z) - On Fairness and Stability: Is Estimator Variance a Friend or a Foe? [6.751310968561177]
We propose a new family of performance measures based on group-wise parity in variance.
We develop and release an open-source library that reconciles uncertainty quantification techniques with fairness analysis.
arXiv Detail & Related papers (2023-02-09T09:35:36Z) - Arbitrariness and Social Prediction: The Confounding Role of Variance in
Fair Classification [31.392067805022414]
Variance in predictions across different trained models is a significant, under-explored source of error in fair binary classification.
In practice, the variance on some data examples is so large that decisions can be effectively arbitrary.
We develop an ensembling algorithm that abstains from classification when a prediction would be arbitrary.
arXiv Detail & Related papers (2023-01-27T06:52:04Z) - Canary in a Coalmine: Better Membership Inference with Ensembled
Adversarial Queries [53.222218035435006]
We use adversarial tools to optimize for queries that are discriminative and diverse.
Our improvements achieve significantly more accurate membership inference than existing methods.
arXiv Detail & Related papers (2022-10-19T17:46:50Z) - The Glass Ceiling of Automatic Evaluation in Natural Language Generation [60.59732704936083]
We take a step back and analyze recent progress by comparing the body of existing automatic metrics and human metrics.
Our extensive statistical analysis reveals surprising findings: automatic metrics -- old and new -- are much more similar to each other than to humans.
arXiv Detail & Related papers (2022-08-31T01:13:46Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Evaluating Metrics for Bias in Word Embeddings [44.14639209617701]
We formalize a bias definition based on the ideas from previous works and derive conditions for bias metrics.
We propose a new metric, SAME, to address the shortcomings of existing metrics and mathematically prove that SAME behaves appropriately.
arXiv Detail & Related papers (2021-11-15T16:07:15Z) - Statistical discrimination in learning agents [64.78141757063142]
Statistical discrimination emerges in agent policies as a function of both the bias in the training population and of agent architecture.
We show that less discrimination emerges with agents that use recurrent neural networks, and when their training environment has less bias.
arXiv Detail & Related papers (2021-10-21T18:28:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.