On Adversarial Bias and the Robustness of Fair Machine Learning
- URL: http://arxiv.org/abs/2006.08669v1
- Date: Mon, 15 Jun 2020 18:17:44 GMT
- Title: On Adversarial Bias and the Robustness of Fair Machine Learning
- Authors: Hongyan Chang, Ta Duy Nguyen, Sasi Kumar Murakonda, Ehsan Kazemi, Reza
Shokri
- Abstract summary: We show that giving the same importance to groups of different sizes and distributions, to counteract the effect of bias in training data, can be in conflict with robustness.
An adversary who can control sampling or labeling for a fraction of training data, can reduce the test accuracy significantly beyond what he can achieve on unconstrained models.
We analyze the robustness of fair machine learning through an empirical evaluation of attacks on multiple algorithms and benchmark datasets.
- Score: 11.584571002297217
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optimizing prediction accuracy can come at the expense of fairness. Towards
minimizing discrimination against a group, fair machine learning algorithms
strive to equalize the behavior of a model across different groups, by imposing
a fairness constraint on models. However, we show that giving the same
importance to groups of different sizes and distributions, to counteract the
effect of bias in training data, can be in conflict with robustness. We analyze
data poisoning attacks against group-based fair machine learning, with the
focus on equalized odds. An adversary who can control sampling or labeling for
a fraction of training data, can reduce the test accuracy significantly beyond
what he can achieve on unconstrained models. Adversarial sampling and
adversarial labeling attacks can also worsen the model's fairness gap on test
data, even though the model satisfies the fairness constraint on training data.
We analyze the robustness of fair machine learning through an empirical
evaluation of attacks on multiple algorithms and benchmark datasets.
Related papers
- Fairness Without Harm: An Influence-Guided Active Sampling Approach [32.173195437797766]
We aim to train models that mitigate group fairness disparity without causing harm to model accuracy.
The current data acquisition methods, such as fair active learning approaches, typically require annotating sensitive attributes.
We propose a tractable active data sampling algorithm that does not rely on training group annotations.
arXiv Detail & Related papers (2024-02-20T07:57:38Z) - On The Impact of Machine Learning Randomness on Group Fairness [11.747264308336012]
We investigate the impact on group fairness of different sources of randomness in training neural networks.
We show that the variance in group fairness measures is rooted in the high volatility of the learning process on under-represented groups.
We show how one can control group-level accuracy, with high efficiency and negligible impact on the model's overall performance, by simply changing the data order for a single epoch.
arXiv Detail & Related papers (2023-07-09T09:36:31Z) - DualFair: Fair Representation Learning at Both Group and Individual
Levels via Contrastive Self-supervision [73.80009454050858]
This work presents a self-supervised model, called DualFair, that can debias sensitive attributes like gender and race from learned representations.
Our model jointly optimize for two fairness criteria - group fairness and counterfactual fairness.
arXiv Detail & Related papers (2023-03-15T07:13:54Z) - Fairness-aware Regression Robust to Adversarial Attacks [46.01773881604595]
We take a first step towards answering the question of how to design fair machine learning algorithms that are robust to adversarial attacks.
For both synthetic data and real-world datasets, numerical results illustrate that the proposed adversarially robust models have better performance on poisoned datasets than other fair machine learning models.
arXiv Detail & Related papers (2022-11-04T18:09:34Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - FairIF: Boosting Fairness in Deep Learning via Influence Functions with
Validation Set Sensitive Attributes [51.02407217197623]
We propose a two-stage training algorithm named FAIRIF.
It minimizes the loss over the reweighted data set where the sample weights are computed.
We show that FAIRIF yields models with better fairness-utility trade-offs against various types of bias.
arXiv Detail & Related papers (2022-01-15T05:14:48Z) - Poisoning Attacks on Fair Machine Learning [13.874416271549523]
We present a framework that seeks to generate poisoning samples to attack both model accuracy and algorithmic fairness.
We develop three online attacks, adversarial sampling, adversarial labeling, and adversarial feature modification.
Our framework enables attackers to flexibly adjust the attack's focus on prediction accuracy or fairness and accurately quantify the impact of each candidate point to both accuracy loss and fairness violation.
arXiv Detail & Related papers (2021-10-17T21:56:14Z) - Fairness in Semi-supervised Learning: Unlabeled Data Help to Reduce
Discrimination [53.3082498402884]
A growing specter in the rise of machine learning is whether the decisions made by machine learning models are fair.
We present a framework of fair semi-supervised learning in the pre-processing phase, including pseudo labeling to predict labels for unlabeled data.
A theoretical decomposition analysis of bias, variance and noise highlights the different sources of discrimination and the impact they have on fairness in semi-supervised learning.
arXiv Detail & Related papers (2020-09-25T05:48:56Z) - Beyond Individual and Group Fairness [90.4666341812857]
We present a new data-driven model of fairness that is guided by the unfairness complaints received by the system.
Our model supports multiple fairness criteria and takes into account their potential incompatibilities.
arXiv Detail & Related papers (2020-08-21T14:14:44Z) - Adversarial Self-Supervised Contrastive Learning [62.17538130778111]
Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions.
We propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples.
We present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data.
arXiv Detail & Related papers (2020-06-13T08:24:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.