FERI: A Multitask-based Fairness Achieving Algorithm with Applications
to Fair Organ Transplantation
- URL: http://arxiv.org/abs/2310.13820v1
- Date: Fri, 20 Oct 2023 21:14:07 GMT
- Title: FERI: A Multitask-based Fairness Achieving Algorithm with Applications
to Fair Organ Transplantation
- Authors: Can Li, Dejian Lai, Xiaoqian Jiang, Kai Zhang
- Abstract summary: We introduce Fairness through the Equitable Rate of Improvement in Multitask Learning (FERI) algorithm for fair predictions of graft failure risk in liver transplant patients.
FERI constrains subgroup loss by balancing learning rates and preventing subgroup dominance in the training process.
Specifically, for gender, FERI reduces the demographic parity disparity by 71.74%, and for the age group, it decreases the equalized odds disparity by 40.46%.
- Score: 16.91239959889591
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Liver transplantation often faces fairness challenges across subgroups
defined by sensitive attributes like age group, gender, and race/ethnicity.
Machine learning models for outcome prediction can introduce additional biases.
To address these, we introduce Fairness through the Equitable Rate of
Improvement in Multitask Learning (FERI) algorithm for fair predictions of
graft failure risk in liver transplant patients. FERI constrains subgroup loss
by balancing learning rates and preventing subgroup dominance in the training
process. Our experiments show that FERI maintains high predictive accuracy with
AUROC and AUPRC comparable to baseline models. More importantly, FERI
demonstrates an ability to improve fairness without sacrificing accuracy.
Specifically, for gender, FERI reduces the demographic parity disparity by
71.74%, and for the age group, it decreases the equalized odds disparity by
40.46%. Therefore, the FERI algorithm advances fairness-aware predictive
modeling in healthcare and provides an invaluable tool for equitable healthcare
systems.
Related papers
- Evaluating Fair Feature Selection in Machine Learning for Healthcare [0.9222623206734782]
We explore algorithmic fairness from the perspective of feature selection.
We evaluate a fair feature selection method that considers equal importance to all demographic groups.
We tested our approach on three publicly available healthcare datasets.
arXiv Detail & Related papers (2024-03-28T06:24:04Z) - Evaluating the Fairness of the MIMIC-IV Dataset and a Baseline
Algorithm: Application to the ICU Length of Stay Prediction [65.268245109828]
This paper uses the MIMIC-IV dataset to examine the fairness and bias in an XGBoost binary classification model predicting the ICU length of stay.
The research reveals class imbalances in the dataset across demographic attributes and employs data preprocessing and feature extraction.
The paper concludes with recommendations for fairness-aware machine learning techniques for mitigating biases and the need for collaborative efforts among healthcare professionals and data scientists.
arXiv Detail & Related papers (2023-12-31T16:01:48Z) - Equal Opportunity of Coverage in Fair Regression [50.76908018786335]
We study fair machine learning (ML) under predictive uncertainty to enable reliable and trustworthy decision-making.
We propose Equal Opportunity of Coverage (EOC) that aims to achieve two properties: (1) coverage rates for different groups with similar outcomes are close, and (2) the coverage rate for the entire population remains at a predetermined level.
arXiv Detail & Related papers (2023-11-03T21:19:59Z) - (Predictable) Performance Bias in Unsupervised Anomaly Detection [3.826262429926079]
Unsupervised anomaly detection (UAD) models promise to aid in the crucial first step of disease detection.
Our study quantified the disparate performance of UAD models against certain demographic subgroups.
arXiv Detail & Related papers (2023-09-25T14:57:43Z) - Auditing ICU Readmission Rates in an Clinical Database: An Analysis of
Risk Factors and Clinical Outcomes [0.0]
This study presents a machine learning pipeline for clinical data classification in the context of a 30-day readmission problem.
The fairness audit uncovers disparities in equal opportunity, predictive parity, false positive rate parity, and false negative rate parity criteria.
The study suggests the need for collaborative efforts among researchers, policymakers, and practitioners to address bias and fairness in artificial intelligence (AI) systems.
arXiv Detail & Related papers (2023-04-12T17:09:38Z) - FairAdaBN: Mitigating unfairness with adaptive batch normalization and
its application to dermatological disease classification [14.589159162086926]
We propose FairAdaBN, which makes batch normalization adaptive to sensitive attribute.
We propose a new metric, named Fairness-Accuracy Trade-off Efficiency (FATE), to compute normalized fairness improvement over accuracy drop.
Experiments on two dermatological datasets show that our proposed method outperforms other methods on fairness criteria and FATE.
arXiv Detail & Related papers (2023-03-15T02:22:07Z) - Chasing Fairness Under Distribution Shift: A Model Weight Perturbation
Approach [72.19525160912943]
We first theoretically demonstrate the inherent connection between distribution shift, data perturbation, and model weight perturbation.
We then analyze the sufficient conditions to guarantee fairness for the target dataset.
Motivated by these sufficient conditions, we propose robust fairness regularization (RFR)
arXiv Detail & Related papers (2023-03-06T17:19:23Z) - Learning Informative Representation for Fairness-aware Multivariate
Time-series Forecasting: A Group-based Perspective [50.093280002375984]
Performance unfairness among variables widely exists in multivariate time series (MTS) forecasting models.
We propose a novel framework, named FairFor, for fairness-aware MTS forecasting.
arXiv Detail & Related papers (2023-01-27T04:54:12Z) - Fair Machine Learning in Healthcare: A Review [90.22219142430146]
We analyze the intersection of fairness in machine learning and healthcare disparities.
We provide a critical review of the associated fairness metrics from a machine learning standpoint.
We propose several new research directions that hold promise for developing ethical and equitable ML applications in healthcare.
arXiv Detail & Related papers (2022-06-29T04:32:10Z) - Assessing Social Determinants-Related Performance Bias of Machine
Learning Models: A case of Hyperchloremia Prediction in ICU Population [6.8473641147443995]
We evaluated four classifiers built to predict Hyperchloremia, a condition that often results from aggressive fluids administration in the ICU population.
We observed that adding social determinants features in addition to the lab-based ones improved model performance on all patients.
We urge future researchers to design models that proactively adjust for potential biases and include subgroup reporting.
arXiv Detail & Related papers (2021-11-18T03:58:50Z) - Estimating and Improving Fairness with Adversarial Learning [65.99330614802388]
We propose an adversarial multi-task training strategy to simultaneously mitigate and detect bias in the deep learning-based medical image analysis system.
Specifically, we propose to add a discrimination module against bias and a critical module that predicts unfairness within the base classification model.
We evaluate our framework on a large-scale public-available skin lesion dataset.
arXiv Detail & Related papers (2021-03-07T03:10:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.