Beyond Internal Data: Bounding and Estimating Fairness from Incomplete Data
- URL: http://arxiv.org/abs/2508.13040v1
- Date: Mon, 18 Aug 2025 15:57:30 GMT
- Title: Beyond Internal Data: Bounding and Estimating Fairness from Incomplete Data
- Authors: Varsha Ramineni, Hossein A. Rahmani, Emine Yilmaz, David Barber,
- Abstract summary: In high-stakes domains such as lending, hiring, and healthcare, ensuring fairness in AI systems is critical.<n>In industry settings, legal and privacy concerns restrict the collection of demographic data required to assess group disparities.<n>Our work seeks to leverage such available separate data to estimate model fairness when complete data is inaccessible.
- Score: 26.037607208689977
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ensuring fairness in AI systems is critical, especially in high-stakes domains such as lending, hiring, and healthcare. This urgency is reflected in emerging global regulations that mandate fairness assessments and independent bias audits. However, procuring the necessary complete data for fairness testing remains a significant challenge. In industry settings, legal and privacy concerns restrict the collection of demographic data required to assess group disparities, and auditors face practical and cultural challenges in gaining access to data. In practice, data relevant for fairness testing is often split across separate sources: internal datasets held by institutions with predictive attributes, and external public datasets such as census data containing protected attributes, each providing only partial, marginal information. Our work seeks to leverage such available separate data to estimate model fairness when complete data is inaccessible. We propose utilising the available separate data to estimate a set of feasible joint distributions and then compute the set plausible fairness metrics. Through simulation and real experiments, we demonstrate that we can derive meaningful bounds on fairness metrics and obtain reliable estimates of the true metric. Our results demonstrate that this approach can serve as a practical and effective solution for fairness testing in real-world settings where access to complete data is restricted.
Related papers
- Beyond Internal Data: Constructing Complete Datasets for Fairness Testing [26.037607208689977]
This work focuses on evaluating classifier fairness when complete datasets including demographics are inaccessible.<n>We propose leveraging separate overlapping datasets to construct complete synthetic data that includes demographic information.<n>We validate the fidelity of the synthetic data by comparing it to real data, and empirically demonstrate that fairness metrics derived from testing on such synthetic data are consistent with those obtained from real data.
arXiv Detail & Related papers (2025-07-24T16:35:42Z) - Testing Fairness with Utility Tradeoffs: A Wasserstein Projection Approach [6.378410364292642]
We propose a statistical hypothesis testing framework that jointly evaluates approximate fairness and utility.<n>Our framework builds on the strong demographic parity criterion and incorporates a utility measure motivated by the potential outcomes framework.<n>We show that the test is computationally tractable, interpretable, broadly applicable across machine learning models, and extendable to more general settings.
arXiv Detail & Related papers (2025-05-16T20:29:06Z) - Targeted Learning for Data Fairness [52.59573714151884]
We expand fairness inference by evaluating fairness in the data generating process itself.<n>We derive estimators demographic parity, equal opportunity, and conditional mutual information.<n>To validate our approach, we perform several simulations and apply our estimators to real data.
arXiv Detail & Related papers (2025-02-06T18:51:28Z) - FairJob: A Real-World Dataset for Fairness in Online Systems [2.3622884172290255]
We introduce a fairness-aware dataset for job recommendations in advertising.
It was collected and prepared to comply with privacy standards and business confidentiality.
Despite being anonymized and including a proxy for a sensitive attribute, our dataset preserves predictive power.
arXiv Detail & Related papers (2024-07-03T12:30:39Z) - Lazy Data Practices Harm Fairness Research [49.02318458244464]
We present a comprehensive analysis of fair ML datasets, demonstrating how unreflective practices hinder the reach and reliability of algorithmic fairness findings.
Our analyses identify three main areas of concern: (1) a textbflack of representation for certain protected attributes in both data and evaluations; (2) the widespread textbf of minorities during data preprocessing; and (3) textbfopaque data processing threatening the generalization of fairness research.
This study underscores the need for a critical reevaluation of data practices in fair ML and offers directions to improve both the sourcing and usage of datasets.
arXiv Detail & Related papers (2024-04-26T09:51:24Z) - Collect, Measure, Repeat: Reliability Factors for Responsible AI Data
Collection [8.12993269922936]
We argue that data collection for AI should be performed in a responsible manner.
We propose a Responsible AI (RAI) methodology designed to guide the data collection with a set of metrics.
arXiv Detail & Related papers (2023-08-22T18:01:27Z) - Auditing and Generating Synthetic Data with Controllable Trust Trade-offs [54.262044436203965]
We introduce a holistic auditing framework that comprehensively evaluates synthetic datasets and AI models.
It focuses on preventing bias and discrimination, ensures fidelity to the source data, assesses utility, robustness, and privacy preservation.
We demonstrate the framework's effectiveness by auditing various generative models across diverse use cases.
arXiv Detail & Related papers (2023-04-21T09:03:18Z) - Data-SUITE: Data-centric identification of in-distribution incongruous
examples [81.21462458089142]
Data-SUITE is a data-centric framework to identify incongruous regions of in-distribution (ID) data.
We empirically validate Data-SUITE's performance and coverage guarantees.
arXiv Detail & Related papers (2022-02-17T18:58:31Z) - Assessing Fairness in the Presence of Missing Data [2.3605348648054463]
We study the problem of estimating fairness in the complete data domain for an arbitrary model evaluated merely using complete cases.
Our work provides the first known theoretical results on fairness guarantee in analysis of incomplete data.
arXiv Detail & Related papers (2021-12-07T17:51:26Z) - Representative & Fair Synthetic Data [68.8204255655161]
We present a framework to incorporate fairness constraints into the self-supervised learning process.
We generate a representative as well as fair version of the UCI Adult census data set.
We consider representative & fair synthetic data a promising future building block to teach algorithms not on historic worlds, but rather on the worlds that we strive to live in.
arXiv Detail & Related papers (2021-04-07T09:19:46Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.