On Explaining Unfairness: An Overview
- URL: http://arxiv.org/abs/2402.10762v1
- Date: Fri, 16 Feb 2024 15:38:00 GMT
- Title: On Explaining Unfairness: An Overview
- Authors: Christos Fragkathoulas, Vasiliki Papanikou, Danae Pla Karidi,
Evaggelia Pitoura
- Abstract summary: Algorithmic fairness and explainability are foundational elements for achieving responsible AI.
We categorize fairness into three types: (a) Explanations to enhance fairness metrics, (b) Explanations to help us understand the causes of (un)fairness, and (c) Explanations to assist us in designing methods for mitigating unfairness.
- Score: 2.0277446818411
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Algorithmic fairness and explainability are foundational elements for
achieving responsible AI. In this paper, we focus on their interplay, a
research area that is recently receiving increasing attention. To this end, we
first present two comprehensive taxonomies, each representing one of the two
complementary fields of study: fairness and explanations. Then, we categorize
explanations for fairness into three types: (a) Explanations to enhance
fairness metrics, (b) Explanations to help us understand the causes of
(un)fairness, and (c) Explanations to assist us in designing methods for
mitigating unfairness. Finally, based on our fairness and explanation
taxonomies, we present undiscovered literature paths revealing gaps that can
serve as valuable insights for future research.
Related papers
- The Odyssey of Commonsense Causality: From Foundational Benchmarks to Cutting-Edge Reasoning [70.16523526957162]
Understanding commonsense causality helps people understand the principles of the real world better.
Despite its significance, a systematic exploration of this topic is notably lacking.
Our work aims to provide a systematic overview, update scholars on recent advancements, and provide a pragmatic guide for beginners.
arXiv Detail & Related papers (2024-06-27T16:30:50Z) - Does Faithfulness Conflict with Plausibility? An Empirical Study in Explainable AI across NLP Tasks [9.979726030996051]
We show that Shapley value and LIME could attain greater faithfulness and plausibility.
Our findings suggest that rather than optimizing for one dimension at the expense of the other, we could seek to optimize explainability algorithms with dual objectives.
arXiv Detail & Related papers (2024-03-29T20:28:42Z) - On the Relationship Between Interpretability and Explainability in Machine Learning [2.828173677501078]
Interpretability and explainability have gained more and more attention in the field of machine learning.
Since both provide information about predictors and their decision process, they are often seen as two independent means for one single end.
This view has led to a dichotomous literature: explainability techniques designed for complex black-box models, or interpretable approaches ignoring the many explainability tools.
arXiv Detail & Related papers (2023-11-20T02:31:08Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Complementary Explanations for Effective In-Context Learning [77.83124315634386]
Large language models (LLMs) have exhibited remarkable capabilities in learning from explanations in prompts.
This work aims to better understand the mechanisms by which explanations are used for in-context learning.
arXiv Detail & Related papers (2022-11-25T04:40:47Z) - What-is and How-to for Fairness in Machine Learning: A Survey,
Reflection, and Perspective [13.124434298120494]
We review and reflect on various fairness notions previously proposed in machine learning literature.
We also consider the long-term impact that is induced by current prediction and decision.
This paper demonstrates the importance of matching the mission (which kind of fairness one would like to enforce) and the means (which spectrum of fairness analysis is of interest) to fulfill the intended purpose.
arXiv Detail & Related papers (2022-06-08T18:05:46Z) - Human Interpretation of Saliency-based Explanation Over Text [65.29015910991261]
We study saliency-based explanations over textual data.
We find that people often mis-interpret the explanations.
We propose a method to adjust saliencies based on model estimates of over- and under-perception.
arXiv Detail & Related papers (2022-01-27T15:20:32Z) - Prompting Contrastive Explanations for Commonsense Reasoning Tasks [74.7346558082693]
Large pretrained language models (PLMs) can achieve near-human performance on commonsense reasoning tasks.
We show how to use these same models to generate human-interpretable evidence.
arXiv Detail & Related papers (2021-06-12T17:06:13Z) - Fairness and Robustness of Contrasting Explanations [9.104557591459283]
We study individual fairness and robustness of contrasting explanations.
We propose to use plausible counterfactuals instead of closest counterfactuals for improving the individual fairness of counterfactual explanations.
arXiv Detail & Related papers (2021-03-03T12:16:06Z) - The Struggles of Feature-Based Explanations: Shapley Values vs. Minimal
Sufficient Subsets [61.66584140190247]
We show that feature-based explanations pose problems even for explaining trivial models.
We show that two popular classes of explainers, Shapley explainers and minimal sufficient subsets explainers, target fundamentally different types of ground-truth explanations.
arXiv Detail & Related papers (2020-09-23T09:45:23Z) - Machine learning fairness notions: Bridging the gap with real-world
applications [4.157415305926584]
Fairness emerged as an important requirement to guarantee that Machine Learning predictive systems do not discriminate against specific individuals or entire sub-populations.
This paper is a survey that illustrates the subtleties between fairness notions through a large number of examples and scenarios.
arXiv Detail & Related papers (2020-06-30T13:01:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.