Fairness-aware Model-agnostic Positive and Unlabeled Learning
- URL: http://arxiv.org/abs/2206.09346v1
- Date: Sun, 19 Jun 2022 08:04:23 GMT
- Title: Fairness-aware Model-agnostic Positive and Unlabeled Learning
- Authors: Ziwei Wu, Jingrui He
- Abstract summary: We propose a fairness-aware Positive and Unlabeled Learning (PUL) method named FairPUL.
For binary classification over individuals from two populations, we aim to achieve similar true positive rates and false positive rates.
Our framework is proven to be statistically consistent in terms of both the classification error and the fairness metric.
- Score: 38.50536380390474
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the increasing application of machine learning in high-stake
decision-making problems, potential algorithmic bias towards people from
certain social groups poses negative impacts on individuals and our society at
large. In the real-world scenario, many such problems involve positive and
unlabeled data such as medical diagnosis, criminal risk assessment and
recommender systems. For instance, in medical diagnosis, only the diagnosed
diseases will be recorded (positive) while others will not (unlabeled). Despite
the large amount of existing work on fairness-aware machine learning in the
(semi-)supervised and unsupervised settings, the fairness issue is largely
under-explored in the aforementioned Positive and Unlabeled Learning (PUL)
context, where it is usually more severe. In this paper, to alleviate this
tension, we propose a fairness-aware PUL method named FairPUL. In particular,
for binary classification over individuals from two populations, we aim to
achieve similar true positive rates and false positive rates in both
populations as our fairness metric. Based on the analysis of the optimal fair
classifier for PUL, we design a model-agnostic post-processing framework,
leveraging both the positive examples and unlabeled ones. Our framework is
proven to be statistically consistent in terms of both the classification error
and the fairness metric. Experiments on the synthetic and real-world data sets
demonstrate that our framework outperforms state-of-the-art in both PUL and
fair classification.
Related papers
- Fairness-enhancing mixed effects deep learning improves fairness on in- and out-of-distribution clustered (non-iid) data [6.596656267996196]
We introduce the Fair Mixed Effects Deep Learning (Fair MEDL) framework.
Fair MEDL quantifies cluster-invariant fixed effects (FE) and cluster-specific random effects (RE)
We incorporate adversarial debiasing to promote fairness across three key metrics: Equalized Odds, Demographic Parity, and Counterfactual Fairness.
arXiv Detail & Related papers (2023-10-04T20:18:45Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Fair Machine Learning in Healthcare: A Review [90.22219142430146]
We analyze the intersection of fairness in machine learning and healthcare disparities.
We provide a critical review of the associated fairness metrics from a machine learning standpoint.
We propose several new research directions that hold promise for developing ethical and equitable ML applications in healthcare.
arXiv Detail & Related papers (2022-06-29T04:32:10Z) - Normalise for Fairness: A Simple Normalisation Technique for Fairness in Regression Machine Learning Problems [46.93320580613236]
We present a simple, yet effective method based on normalisation (FaiReg) for regression problems.
We compare it with two standard methods for fairness, namely data balancing and adversarial training.
The results show the superior performance of diminishing the effects of unfairness better than data balancing.
arXiv Detail & Related papers (2022-02-02T12:26:25Z) - Blackbox Post-Processing for Multiclass Fairness [1.5305403478254664]
We consider modifying the predictions of a blackbox machine learning classifier in order to achieve fairness in a multiclass setting.
We explore when our approach produces both fair and accurate predictions through systematic synthetic experiments.
We find that overall, our approach produces minor drops in accuracy and enforces fairness when the number of individuals in the dataset is high.
arXiv Detail & Related papers (2022-01-12T13:21:20Z) - Fairness-aware Class Imbalanced Learning [57.45784950421179]
We evaluate long-tail learning methods for tweet sentiment and occupation classification.
We extend a margin-loss based approach with methods to enforce fairness.
arXiv Detail & Related papers (2021-09-21T22:16:30Z) - Estimating and Improving Fairness with Adversarial Learning [65.99330614802388]
We propose an adversarial multi-task training strategy to simultaneously mitigate and detect bias in the deep learning-based medical image analysis system.
Specifically, we propose to add a discrimination module against bias and a critical module that predicts unfairness within the base classification model.
We evaluate our framework on a large-scale public-available skin lesion dataset.
arXiv Detail & Related papers (2021-03-07T03:10:32Z) - Fairness in Semi-supervised Learning: Unlabeled Data Help to Reduce
Discrimination [53.3082498402884]
A growing specter in the rise of machine learning is whether the decisions made by machine learning models are fair.
We present a framework of fair semi-supervised learning in the pre-processing phase, including pseudo labeling to predict labels for unlabeled data.
A theoretical decomposition analysis of bias, variance and noise highlights the different sources of discrimination and the impact they have on fairness in semi-supervised learning.
arXiv Detail & Related papers (2020-09-25T05:48:56Z) - Recovering from Biased Data: Can Fairness Constraints Improve Accuracy? [11.435833538081557]
Empirical Risk Minimization (ERM) may produce a classifier that not only is biased but also has suboptimal accuracy on the true data distribution.
We examine the ability of fairness-constrained ERM to correct this problem.
We also consider other recovery methods including reweighting the training data, Equalized Odds, and Demographic Parity.
arXiv Detail & Related papers (2019-12-02T22:00:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.