Practical Guide for Causal Pathways and Sub-group Disparity Analysis
- URL: http://arxiv.org/abs/2407.02702v3
- Date: Wed, 7 Aug 2024 01:35:58 GMT
- Title: Practical Guide for Causal Pathways and Sub-group Disparity Analysis
- Authors: Farnaz Kohankhaki, Shaina Raza, Oluwanifemi Bamgbose, Deval Pandya, Elham Dolatabadi,
- Abstract summary: We use causal disparity analysis to quantify and examine the causal interplay between sensitive attributes and outcomes.
Our two-step investigation focuses on datasets where race serves as the sensitive attribute.
We demonstrate that the sub-groups identified by our approach to be affected the most by disparities are the ones with the largest ML classification errors.
- Score: 1.8974791957167259
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this study, we introduce the application of causal disparity analysis to unveil intricate relationships and causal pathways between sensitive attributes and the targeted outcomes within real-world observational data. Our methodology involves employing causal decomposition analysis to quantify and examine the causal interplay between sensitive attributes and outcomes. We also emphasize the significance of integrating heterogeneity assessment in causal disparity analysis to gain deeper insights into the impact of sensitive attributes within specific sub-groups on outcomes. Our two-step investigation focuses on datasets where race serves as the sensitive attribute. The results on two datasets indicate the benefit of leveraging causal analysis and heterogeneity assessment not only for quantifying biases in the data but also for disentangling their influences on outcomes. We demonstrate that the sub-groups identified by our approach to be affected the most by disparities are the ones with the largest ML classification errors. We also show that grouping the data only based on a sensitive attribute is not enough, and through these analyses, we can find sub-groups that are directly affected by disparities. We hope that our findings will encourage the adoption of such methodologies in future ethical AI practices and bias audits, fostering a more equitable and fair technological landscape.
Related papers
- Unsupervised Pairwise Causal Discovery on Heterogeneous Data using Mutual Information Measures [49.1574468325115]
Causal Discovery is a technique that tackles the challenge by analyzing the statistical properties of the constituent variables.
We question the current (possibly misleading) baseline results on the basis that they were obtained through supervised learning.
In consequence, we approach this problem in an unsupervised way, using robust Mutual Information measures.
arXiv Detail & Related papers (2024-08-01T09:11:08Z) - Bridging the Gap: Protocol Towards Fair and Consistent Affect Analysis [24.737468736951374]
The increasing integration of machine learning algorithms in daily life underscores the critical need for fairness and equity in their deployment.
Existing databases and methodologies lack uniformity, leading to biased evaluations.
This work addresses these issues by analyzing six affective databases, annotating demographic attributes, and proposing a common protocol for database partitioning.
arXiv Detail & Related papers (2024-05-10T22:40:01Z) - A Neural Framework for Generalized Causal Sensitivity Analysis [78.71545648682705]
We propose NeuralCSA, a neural framework for causal sensitivity analysis.
We provide theoretical guarantees that NeuralCSA is able to infer valid bounds on the causal query of interest.
arXiv Detail & Related papers (2023-11-27T17:40:02Z) - The Role of Subgroup Separability in Group-Fair Medical Image
Classification [18.29079361470428]
We find a relationship between subgroup separability, subgroup disparities, and performance degradation when models are trained on data with systematic bias such as underdiagnosis.
Our findings shed new light on the question of how models become biased, providing important insights for the development of fair medical imaging AI.
arXiv Detail & Related papers (2023-07-06T06:06:47Z) - A Causal Framework for Decomposing Spurious Variations [68.12191782657437]
We develop tools for decomposing spurious variations in Markovian and Semi-Markovian models.
We prove the first results that allow a non-parametric decomposition of spurious effects.
The described approach has several applications, ranging from explainable and fair AI to questions in epidemiology and medicine.
arXiv Detail & Related papers (2023-06-08T09:40:28Z) - Sharp Bounds for Generalized Causal Sensitivity Analysis [30.77874108094485]
We propose a unified framework for causal sensitivity analysis under unobserved confounding.
This includes (conditional) average treatment effects, effects for mediation analysis and path analysis, and distributional effects.
Our bounds for (conditional) average treatment effects coincide with recent optimality results for causal sensitivity analysis.
arXiv Detail & Related papers (2023-05-26T14:44:32Z) - Towards Unbiased Visual Emotion Recognition via Causal Intervention [63.74095927462]
We propose a novel Emotion Recognition Network (IERN) to alleviate the negative effects brought by the dataset bias.
A series of designed tests validate the effectiveness of IERN, and experiments on three emotion benchmarks demonstrate that IERN outperforms other state-of-the-art approaches.
arXiv Detail & Related papers (2021-07-26T10:40:59Z) - Through the Data Management Lens: Experimental Analysis and Evaluation
of Fair Classification [75.49600684537117]
Data management research is showing an increasing presence and interest in topics related to data and algorithmic fairness.
We contribute a broad analysis of 13 fair classification approaches and additional variants, over their correctness, fairness, efficiency, scalability, and stability.
Our analysis highlights novel insights on the impact of different metrics and high-level approach characteristics on different aspects of performance.
arXiv Detail & Related papers (2021-01-18T22:55:40Z) - Targeted VAE: Variational and Targeted Learning for Causal Inference [39.351088248776435]
Undertaking causal inference with observational data is incredibly useful across a wide range of tasks.
There are two significant challenges associated with undertaking causal inference using observational data.
We address these two challenges by combining structured inference and targeted learning.
arXiv Detail & Related papers (2020-09-28T16:55:24Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z) - Interpreting Deep Models through the Lens of Data [5.174367472975529]
This paper presents an in-depth analysis of the methods which attempt to identify the influence of these data points on the resulting classifier.
We show that some interpretability methods can detect mislabels better than using a random approach, however, the sample selection based on the training loss showed a superior performance.
arXiv Detail & Related papers (2020-05-05T07:59:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.