An Empirical Characterization of Fair Machine Learning For Clinical Risk
Prediction
- URL: http://arxiv.org/abs/2007.10306v3
- Date: Tue, 15 Jun 2021 15:28:53 GMT
- Title: An Empirical Characterization of Fair Machine Learning For Clinical Risk
Prediction
- Authors: Stephen R. Pfohl, Agata Foryciarz, Nigam H. Shah
- Abstract summary: The use of machine learning to guide clinical decision making has the potential to worsen existing health disparities.
Several recent works frame the problem as that of algorithmic fairness, a framework that has attracted considerable attention and criticism.
We conduct an empirical study to characterize the impact of penalizing group fairness violations on an array of measures of model performance and group fairness.
- Score: 7.945729033499554
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The use of machine learning to guide clinical decision making has the
potential to worsen existing health disparities. Several recent works frame the
problem as that of algorithmic fairness, a framework that has attracted
considerable attention and criticism. However, the appropriateness of this
framework is unclear due to both ethical as well as technical considerations,
the latter of which include trade-offs between measures of fairness and model
performance that are not well-understood for predictive models of clinical
outcomes. To inform the ongoing debate, we conduct an empirical study to
characterize the impact of penalizing group fairness violations on an array of
measures of model performance and group fairness. We repeat the analyses across
multiple observational healthcare databases, clinical outcomes, and sensitive
attributes. We find that procedures that penalize differences between the
distributions of predictions across groups induce nearly-universal degradation
of multiple performance metrics within groups. On examining the secondary
impact of these procedures, we observe heterogeneity of the effect of these
procedures on measures of fairness in calibration and ranking across
experimental conditions. Beyond the reported trade-offs, we emphasize that
analyses of algorithmic fairness in healthcare lack the contextual grounding
and causal awareness necessary to reason about the mechanisms that lead to
health disparities, as well as about the potential of algorithmic fairness
methods to counteract those mechanisms. In light of these limitations, we
encourage researchers building predictive models for clinical use to step
outside the algorithmic fairness frame and engage critically with the broader
sociotechnical context surrounding the use of machine learning in healthcare.
Related papers
- Auditing ICU Readmission Rates in an Clinical Database: An Analysis of
Risk Factors and Clinical Outcomes [0.0]
This study presents a machine learning pipeline for clinical data classification in the context of a 30-day readmission problem.
The fairness audit uncovers disparities in equal opportunity, predictive parity, false positive rate parity, and false negative rate parity criteria.
The study suggests the need for collaborative efforts among researchers, policymakers, and practitioners to address bias and fairness in artificial intelligence (AI) systems.
arXiv Detail & Related papers (2023-04-12T17:09:38Z) - Towards Fair Patient-Trial Matching via Patient-Criterion Level Fairness
Constraint [50.35075018041199]
This work proposes a fair patient-trial matching framework by generating a patient-criterion level fairness constraint.
The experimental results on real-world patient-trial and patient-criterion matching tasks demonstrate that the proposed framework can successfully alleviate the predictions that tend to be biased.
arXiv Detail & Related papers (2023-03-24T03:59:19Z) - Fair Machine Learning in Healthcare: A Review [90.22219142430146]
We analyze the intersection of fairness in machine learning and healthcare disparities.
We provide a critical review of the associated fairness metrics from a machine learning standpoint.
We propose several new research directions that hold promise for developing ethical and equitable ML applications in healthcare.
arXiv Detail & Related papers (2022-06-29T04:32:10Z) - Benchmarking Heterogeneous Treatment Effect Models through the Lens of
Interpretability [82.29775890542967]
Estimating personalized effects of treatments is a complex, yet pervasive problem.
Recent developments in the machine learning literature on heterogeneous treatment effect estimation gave rise to many sophisticated, but opaque, tools.
We use post-hoc feature importance methods to identify features that influence the model's predictions.
arXiv Detail & Related papers (2022-06-16T17:59:05Z) - Net benefit, calibration, threshold selection, and training objectives
for algorithmic fairness in healthcare [7.500205995624894]
We evaluate the interplay between measures of model performance, fairness, and the expected utility of decision-making.
We conduct an empirical case-study via development of models to estimate the ten-year risk of atherosclerotic cardiovascular disease.
arXiv Detail & Related papers (2022-02-03T23:23:05Z) - SurvITE: Learning Heterogeneous Treatment Effects from Time-to-Event
Data [83.50281440043241]
We study the problem of inferring heterogeneous treatment effects from time-to-event data.
We propose a novel deep learning method for treatment-specific hazard estimation based on balancing representations.
arXiv Detail & Related papers (2021-10-26T20:13:17Z) - Measuring Fairness Under Unawareness of Sensitive Attributes: A
Quantification-Based Approach [131.20444904674494]
We tackle the problem of measuring group fairness under unawareness of sensitive attributes.
We show that quantification approaches are particularly suited to tackle the fairness-under-unawareness problem.
arXiv Detail & Related papers (2021-09-17T13:45:46Z) - Fair Conformal Predictors for Applications in Medical Imaging [4.236384785644418]
Conformal methods can complement deep learning models by providing both clinically intuitive way of expressing model uncertainty.
We conduct experiments with a mammographic breast density and dermatology photography datasets to demonstrate the utility of conformal predictions.
We find that conformal predictors can be used to equalize coverage with respect to patient demographics such as race and skin tone.
arXiv Detail & Related papers (2021-09-09T16:31:10Z) - Uncertainty estimation for classification and risk prediction on medical
tabular data [0.0]
This work advances the understanding of uncertainty estimation for classification and risk prediction on medical data.
In a data-scarce field such as healthcare, the ability to measure the uncertainty of a model's prediction could potentially lead to improved effectiveness of decision support tools.
arXiv Detail & Related papers (2020-04-13T08:46:41Z) - Generalization Bounds and Representation Learning for Estimation of
Potential Outcomes and Causal Effects [61.03579766573421]
We study estimation of individual-level causal effects, such as a single patient's response to alternative medication.
We devise representation learning algorithms that minimize our bound, by regularizing the representation's induced treatment group distance.
We extend these algorithms to simultaneously learn a weighted representation to further reduce treatment group distances.
arXiv Detail & Related papers (2020-01-21T10:16:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.