M$^3$Fair: Mitigating Bias in Healthcare Data through Multi-Level and
Multi-Sensitive-Attribute Reweighting Method
- URL: http://arxiv.org/abs/2306.04118v1
- Date: Wed, 7 Jun 2023 03:20:44 GMT
- Title: M$^3$Fair: Mitigating Bias in Healthcare Data through Multi-Level and
Multi-Sensitive-Attribute Reweighting Method
- Authors: Yinghao Zhu, Jingkun An, Enshen Zhou, Lu An, Junyi Gao, Hao Li, Haoran
Feng, Bo Hou, Wen Tang, Chengwei Pan, Liantao Ma
- Abstract summary: We propose M3Fair, a multi-level and multi-sensitive-attribute reweighting method by extending the RW method to multiple sensitive attributes at multiple levels.
Our experiments on real-world datasets show that the approach is effective, straightforward, and generalizable in addressing the healthcare fairness issues.
- Score: 13.253174531040106
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In the data-driven artificial intelligence paradigm, models heavily rely on
large amounts of training data. However, factors like sampling distribution
imbalance can lead to issues of bias and unfairness in healthcare data.
Sensitive attributes, such as race, gender, age, and medical condition, are
characteristics of individuals that are commonly associated with discrimination
or bias. In healthcare AI, these attributes can play a significant role in
determining the quality of care that individuals receive. For example, minority
groups often receive fewer procedures and poorer-quality medical care than
white individuals in US. Therefore, detecting and mitigating bias in data is
crucial to enhancing health equity. Bias mitigation methods include
pre-processing, in-processing, and post-processing. Among them, Reweighting
(RW) is a widely used pre-processing method that performs well in balancing
machine learning performance and fairness performance. RW adjusts the weights
for samples within each (group, label) combination, where these weights are
utilized in loss functions. However, RW is limited to considering only a single
sensitive attribute when mitigating bias and assumes that each sensitive
attribute is equally important. This may result in potential inaccuracies when
addressing intersectional bias. To address these limitations, we propose
M3Fair, a multi-level and multi-sensitive-attribute reweighting method by
extending the RW method to multiple sensitive attributes at multiple levels.
Our experiments on real-world datasets show that the approach is effective,
straightforward, and generalizable in addressing the healthcare fairness
issues.
Related papers
- Evaluating Fair Feature Selection in Machine Learning for Healthcare [0.9222623206734782]
We explore algorithmic fairness from the perspective of feature selection.
We evaluate a fair feature selection method that considers equal importance to all demographic groups.
We tested our approach on three publicly available healthcare datasets.
arXiv Detail & Related papers (2024-03-28T06:24:04Z) - Achieve Fairness without Demographics for Dermatological Disease
Diagnosis [17.792332189055223]
We propose a method enabling fair predictions for sensitive attributes during the testing phase without using such information during training.
Inspired by prior work highlighting the impact of feature entanglement on fairness, we enhance the model features by capturing the features related to the sensitive and target attributes.
This ensures that the model can only classify based on the features related to the target attribute without relying on features associated with sensitive attributes.
arXiv Detail & Related papers (2024-01-16T02:49:52Z) - MCRAGE: Synthetic Healthcare Data for Fairness [3.0089659534785853]
We propose Minority Class Rebalancing through Augmentation by Generative modeling (MCRAGE) to augment imbalanced datasets.
MCRAGE involves training a Denoising Diffusion Probabilistic Model (CDDPM) capable of generating high-quality synthetic EHR samples from underrepresented classes.
We use this synthetic data to augment the existing imbalanced dataset, resulting in a more balanced distribution across all classes.
arXiv Detail & Related papers (2023-10-27T19:02:22Z) - Fast Model Debias with Machine Unlearning [54.32026474971696]
Deep neural networks might behave in a biased manner in many real-world scenarios.
Existing debiasing methods suffer from high costs in bias labeling or model re-training.
We propose a fast model debiasing framework (FMD) which offers an efficient approach to identify, evaluate and remove biases.
arXiv Detail & Related papers (2023-10-19T08:10:57Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Semi-FairVAE: Semi-supervised Fair Representation Learning with
Adversarial Variational Autoencoder [92.67156911466397]
We propose a semi-supervised fair representation learning approach based on adversarial variational autoencoder.
We use a bias-aware model to capture inherent bias information on sensitive attribute.
We also use a bias-free model to learn debiased fair representations by using adversarial learning to remove bias information from them.
arXiv Detail & Related papers (2022-04-01T15:57:47Z) - To Impute or not to Impute? -- Missing Data in Treatment Effect
Estimation [84.76186111434818]
We identify a new missingness mechanism, which we term mixed confounded missingness (MCM), where some missingness determines treatment selection and other missingness is determined by treatment selection.
We show that naively imputing all data leads to poor performing treatment effects models, as the act of imputation effectively removes information necessary to provide unbiased estimates.
Our solution is selective imputation, where we use insights from MCM to inform precisely which variables should be imputed and which should not.
arXiv Detail & Related papers (2022-02-04T12:08:31Z) - Normalise for Fairness: A Simple Normalisation Technique for Fairness in Regression Machine Learning Problems [46.93320580613236]
We present a simple, yet effective method based on normalisation (FaiReg) for regression problems.
We compare it with two standard methods for fairness, namely data balancing and adversarial training.
The results show the superior performance of diminishing the effects of unfairness better than data balancing.
arXiv Detail & Related papers (2022-02-02T12:26:25Z) - MultiFair: Multi-Group Fairness in Machine Learning [52.24956510371455]
We study multi-group fairness in machine learning (MultiFair)
We propose a generic end-to-end algorithmic framework to solve it.
Our proposed framework is generalizable to many different settings.
arXiv Detail & Related papers (2021-05-24T02:30:22Z) - Estimating and Improving Fairness with Adversarial Learning [65.99330614802388]
We propose an adversarial multi-task training strategy to simultaneously mitigate and detect bias in the deep learning-based medical image analysis system.
Specifically, we propose to add a discrimination module against bias and a critical module that predicts unfairness within the base classification model.
We evaluate our framework on a large-scale public-available skin lesion dataset.
arXiv Detail & Related papers (2021-03-07T03:10:32Z) - Recovering from Biased Data: Can Fairness Constraints Improve Accuracy? [11.435833538081557]
Empirical Risk Minimization (ERM) may produce a classifier that not only is biased but also has suboptimal accuracy on the true data distribution.
We examine the ability of fairness-constrained ERM to correct this problem.
We also consider other recovery methods including reweighting the training data, Equalized Odds, and Demographic Parity.
arXiv Detail & Related papers (2019-12-02T22:00:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.