Are Two Heads the Same as One? Identifying Disparate Treatment in Fair
Neural Networks
- URL: http://arxiv.org/abs/2204.04440v1
- Date: Sat, 9 Apr 2022 10:07:02 GMT
- Title: Are Two Heads the Same as One? Identifying Disparate Treatment in Fair
Neural Networks
- Authors: Michael Lohaus, Matth\"aus Kleindessner, Krishnaram Kenthapadi,
Francesco Locatello, Chris Russell
- Abstract summary: We show that deep neural networks that satisfy demographic parity do so through a form of race or gender awareness.
We propose a simple two-stage solution for enforcing fairness.
- Score: 38.14455334700671
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We show that deep neural networks that satisfy demographic parity do so
through a form of race or gender awareness, and that the more we force a
network to be fair, the more accurately we can recover race or gender from the
internal state of the network. Based on this observation, we propose a simple
two-stage solution for enforcing fairness. First, we train a two-headed network
to predict the protected attribute (such as race or gender) alongside the
original task, and second, we enforce demographic parity by taking a weighted
sum of the heads. In the end, this approach creates a single-headed network
with the same backbone architecture as the original network. Our approach has
near identical performance compared to existing regularization-based or
preprocessing methods, but has greater stability and higher accuracy where near
exact demographic parity is required. To cement the relationship between these
two approaches, we show that an unfair and optimally accurate classifier can be
recovered by taking a weighted sum of a fair classifier and a classifier
predicting the protected attribute. We use this to argue that both the fairness
approaches and our explicit formulation demonstrate disparate treatment and
that, consequentially, they are likely to be unlawful in a wide range of
scenarios under the US law.
Related papers
- Counterfactual Fairness for Predictions using Generative Adversarial
Networks [28.65556399421874]
We develop a novel deep neural network called Generative Counterfactual Fairness Network (GCFN) for making predictions under counterfactual fairness.
Our method is mathematically guaranteed to ensure the notion of counterfactual fairness.
arXiv Detail & Related papers (2023-10-26T17:58:39Z) - Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - DualFair: Fair Representation Learning at Both Group and Individual
Levels via Contrastive Self-supervision [73.80009454050858]
This work presents a self-supervised model, called DualFair, that can debias sensitive attributes like gender and race from learned representations.
Our model jointly optimize for two fairness criteria - group fairness and counterfactual fairness.
arXiv Detail & Related papers (2023-03-15T07:13:54Z) - A Differentiable Distance Approximation for Fairer Image Classification [31.471917430653626]
We propose a differentiable approximation of the variance of demographics, a metric that can be used to measure the bias, or unfairness, in an AI model.
Our approximation can be optimised alongside the regular training objective which eliminates the need for any extra models during training.
We demonstrate that our approach improves the fairness of AI models in varied task and dataset scenarios.
arXiv Detail & Related papers (2022-10-09T23:02:18Z) - CertiFair: A Framework for Certified Global Fairness of Neural Networks [1.4620086904601473]
Individual Fairness suggests that similar individuals with respect to a certain task are to be treated similarly by a Neural Network (NN) model.
We construct a verifier which checks whether the fairness property holds for a given NN in a classification task.
We then provide provable bounds on the fairness of the resulting NN.
arXiv Detail & Related papers (2022-05-20T02:08:47Z) - On Disentangled and Locally Fair Representations [95.6635227371479]
We study the problem of performing classification in a manner that is fair for sensitive groups, such as race and gender.
We learn a locally fair representation, such that, under the learned representation, the neighborhood of each sample is balanced in terms of the sensitive attribute.
arXiv Detail & Related papers (2022-05-05T14:26:50Z) - Probabilistic Verification of Neural Networks Against Group Fairness [21.158245095699456]
We propose an approach to formally verify neural networks against fairness.
Our method is built upon an approach for learning Markov Chains from a user-provided neural network.
We demonstrate that with our analysis results, the neural weights can be optimized to improve fairness.
arXiv Detail & Related papers (2021-07-18T04:34:31Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - FaiR-N: Fair and Robust Neural Networks for Structured Data [10.14835182649819]
We present a novel formulation for training neural networks that considers the distance of data points to the decision boundary.
We show that training with this loss yields more fair and robust neural networks with similar accuracies to models trained without it.
arXiv Detail & Related papers (2020-10-13T01:53:15Z) - Fairness Through Robustness: Investigating Robustness Disparity in Deep
Learning [61.93730166203915]
We argue that traditional notions of fairness are not sufficient when the model is vulnerable to adversarial attacks.
We show that measuring robustness bias is a challenging task for DNNs and propose two methods to measure this form of bias.
arXiv Detail & Related papers (2020-06-17T22:22:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.