Estimating and Implementing Conventional Fairness Metrics With
Probabilistic Protected Features
- URL: http://arxiv.org/abs/2310.01679v1
- Date: Mon, 2 Oct 2023 22:30:25 GMT
- Title: Estimating and Implementing Conventional Fairness Metrics With
Probabilistic Protected Features
- Authors: Hadi Elzayn, Emily Black, Patrick Vossler, Nathanael Jo, Jacob Goldin,
Daniel E. Ho
- Abstract summary: We develop methods for measuring and reducing violations in a setting with limited attribute labels.
We show our measurement method can bound the true disparity up to 5.5x tighter than previous methods in these applications.
- Score: 7.457585597068654
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The vast majority of techniques to train fair models require access to the
protected attribute (e.g., race, gender), either at train time or in
production. However, in many important applications this protected attribute is
largely unavailable. In this paper, we develop methods for measuring and
reducing fairness violations in a setting with limited access to protected
attribute labels. Specifically, we assume access to protected attribute labels
on a small subset of the dataset of interest, but only probabilistic estimates
of protected attribute labels (e.g., via Bayesian Improved Surname Geocoding)
for the rest of the dataset. With this setting in mind, we propose a method to
estimate bounds on common fairness metrics for an existing model, as well as a
method for training a model to limit fairness violations by solving a
constrained non-convex optimization problem. Unlike similar existing
approaches, our methods take advantage of contextual information --
specifically, the relationships between a model's predictions and the
probabilistic prediction of protected attributes, given the true protected
attribute, and vice versa -- to provide tighter bounds on the true disparity.
We provide an empirical illustration of our methods using voting data. First,
we show our measurement method can bound the true disparity up to 5.5x tighter
than previous methods in these applications. Then, we demonstrate that our
training technique effectively reduces disparity while incurring lesser
fairness-accuracy trade-offs than other fair optimization methods with limited
access to protected attributes.
Related papers
- Pseudo-Probability Unlearning: Towards Efficient and Privacy-Preserving Machine Unlearning [59.29849532966454]
We propose PseudoProbability Unlearning (PPU), a novel method that enables models to forget data to adhere to privacy-preserving manner.
Our method achieves over 20% improvements in forgetting error compared to the state-of-the-art.
arXiv Detail & Related papers (2024-11-04T21:27:06Z) - Fairness Without Harm: An Influence-Guided Active Sampling Approach [32.173195437797766]
We aim to train models that mitigate group fairness disparity without causing harm to model accuracy.
The current data acquisition methods, such as fair active learning approaches, typically require annotating sensitive attributes.
We propose a tractable active data sampling algorithm that does not rely on training group annotations.
arXiv Detail & Related papers (2024-02-20T07:57:38Z) - Theoretically Principled Federated Learning for Balancing Privacy and
Utility [61.03993520243198]
We propose a general learning framework for the protection mechanisms that protects privacy via distorting model parameters.
It can achieve personalized utility-privacy trade-off for each model parameter, on each client, at each communication round in federated learning.
arXiv Detail & Related papers (2023-05-24T13:44:02Z) - Can Querying for Bias Leak Protected Attributes? Achieving Privacy With
Smooth Sensitivity [14.564919048514897]
Existing regulations prohibit model developers from accessing protected attributes.
We show that simply querying for fairness metrics can leak the protected attributes of individuals to the model developers.
We propose Attribute-Conceal, a novel technique that achieves differential privacy by calibrating noise to the smooth sensitivity of our bias query.
arXiv Detail & Related papers (2022-11-03T20:44:48Z) - Fair mapping [0.0]
We propose a novel pre-processing method based on the transformation of the distribution of protected groups onto a chosen target one.
We leverage on the recent works of the Wasserstein GAN and AttGAN frameworks to achieve the optimal transport of data points.
Our proposed approach, preserves the interpretability of data and can be used without defining exactly the sensitive groups.
arXiv Detail & Related papers (2022-09-01T17:31:27Z) - A Prototype-Oriented Framework for Unsupervised Domain Adaptation [52.25537670028037]
We provide a memory and computation-efficient probabilistic framework to extract class prototypes and align the target features with them.
We demonstrate the general applicability of our method on a wide range of scenarios, including single-source, multi-source, class-imbalance, and source-private domain adaptation.
arXiv Detail & Related papers (2021-10-22T19:23:22Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - Measuring Fairness Under Unawareness of Sensitive Attributes: A
Quantification-Based Approach [131.20444904674494]
We tackle the problem of measuring group fairness under unawareness of sensitive attributes.
We show that quantification approaches are particularly suited to tackle the fairness-under-unawareness problem.
arXiv Detail & Related papers (2021-09-17T13:45:46Z) - Estimation of Fair Ranking Metrics with Incomplete Judgments [70.37717864975387]
We propose a sampling strategy and estimation technique for four fair ranking metrics.
We formulate a robust and unbiased estimator which can operate even with very limited number of labeled items.
arXiv Detail & Related papers (2021-08-11T10:57:00Z) - Fair Classification with Adversarial Perturbations [35.030329189029246]
We study fair classification in the presence of an omniscient adversary that, given an $eta$, is allowed to choose an arbitrary $eta$-fraction of the training samples and arbitrarily perturb their protected attributes.
Our main contribution is an optimization framework to learn fair classifiers in this adversarial setting that comes with provable guarantees on accuracy and fairness.
We prove near-tightness of our framework's guarantees for natural hypothesis classes: no algorithm can have significantly better accuracy and any algorithm with better fairness must have lower accuracy.
arXiv Detail & Related papers (2021-06-10T17:56:59Z) - Mitigating Bias in Set Selection with Noisy Protected Attributes [16.882719401742175]
We show that in the presence of noisy protected attributes, in attempting to increase fairness without considering noise, one can, in fact, decrease the fairness of the result!
We formulate a denoised'' selection problem which functions for a large class of fairness metrics.
Our empirical results show that this approach can produce subsets which significantly improve the fairness metrics despite the presence of noisy protected attributes.
arXiv Detail & Related papers (2020-11-09T06:45:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.