Fairness and Explainability: Bridging the Gap Towards Fair Model
Explanations
- URL: http://arxiv.org/abs/2212.03840v1
- Date: Wed, 7 Dec 2022 18:35:54 GMT
- Title: Fairness and Explainability: Bridging the Gap Towards Fair Model
Explanations
- Authors: Yuying Zhao, Yu Wang, Tyler Derr
- Abstract summary: We bridge the gap between fairness and explainability by presenting a novel perspective of procedure-oriented fairness based on explanations.
We propose a Comprehensive Fairness Algorithm (CFA), which simultaneously fulfills multiple objectives - improving traditional fairness, satisfying explanation fairness, and maintaining the utility performance.
- Score: 12.248793742165278
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While machine learning models have achieved unprecedented success in
real-world applications, they might make biased/unfair decisions for specific
demographic groups and hence result in discriminative outcomes. Although
research efforts have been devoted to measuring and mitigating bias, they
mainly study bias from the result-oriented perspective while neglecting the
bias encoded in the decision-making procedure. This results in their inability
to capture procedure-oriented bias, which therefore limits the ability to have
a fully debiasing method. Fortunately, with the rapid development of
explainable machine learning, explanations for predictions are now available to
gain insights into the procedure. In this work, we bridge the gap between
fairness and explainability by presenting a novel perspective of
procedure-oriented fairness based on explanations. We identify the
procedure-based bias by measuring the gap of explanation quality between
different groups with Ratio-based and Value-based Explanation Fairness. The new
metrics further motivate us to design an optimization objective to mitigate the
procedure-based bias where we observe that it will also mitigate bias from the
prediction. Based on our designed optimization objective, we propose a
Comprehensive Fairness Algorithm (CFA), which simultaneously fulfills multiple
objectives - improving traditional fairness, satisfying explanation fairness,
and maintaining the utility performance. Extensive experiments on real-world
datasets demonstrate the effectiveness of our proposed CFA and highlight the
importance of considering fairness from the explainability perspective. Our
code is publicly available at
https://github.com/YuyingZhao/FairExplanations-CFA .
Related papers
- Achieving Fairness in Predictive Process Analytics via Adversarial Learning [50.31323204077591]
This paper addresses the challenge of integrating a debiasing phase into predictive business process analytics.
Our framework leverages on adversial debiasing is evaluated on four case studies, showing a significant reduction in the contribution of biased variables to the predicted value.
arXiv Detail & Related papers (2024-10-03T15:56:03Z) - Editable Fairness: Fine-Grained Bias Mitigation in Language Models [52.66450426729818]
We propose a novel debiasing approach, Fairness Stamp (FAST), which enables fine-grained calibration of individual social biases.
FAST surpasses state-of-the-art baselines with superior debiasing performance.
This highlights the potential of fine-grained debiasing strategies to achieve fairness in large language models.
arXiv Detail & Related papers (2024-08-07T17:14:58Z) - Large Language Model Bias Mitigation from the Perspective of Knowledge Editing [12.855975783223236]
We propose a novel debiasing method, Fairness Stamp (FAST), which enables editable fairness through fine-grained calibration on individual biased knowledge.
FAST surpasses state-of-the-art baselines with remarkable debiasing performance while not hampering overall model capability for knowledge preservation.
arXiv Detail & Related papers (2024-05-15T13:44:13Z) - Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias in Factual Knowledge Extraction [56.17020601803071]
Recent research shows that pre-trained language models (PLMs) suffer from "prompt bias" in factual knowledge extraction.
This paper aims to improve the reliability of existing benchmarks by thoroughly investigating and mitigating prompt bias.
arXiv Detail & Related papers (2024-03-15T02:04:35Z) - Fairness Explainability using Optimal Transport with Applications in
Image Classification [0.46040036610482665]
We propose a comprehensive approach to uncover the causes of discrimination in Machine Learning applications.
We leverage Wasserstein barycenters to achieve fair predictions and introduce an extension to pinpoint bias-associated regions.
This allows us to derive a cohesive system which uses the enforced fairness to measure each features influence emphon the bias.
arXiv Detail & Related papers (2023-08-22T00:10:23Z) - Learning for Counterfactual Fairness from Observational Data [62.43249746968616]
Fairness-aware machine learning aims to eliminate biases of learning models against certain subgroups described by certain protected (sensitive) attributes such as race, gender, and age.
A prerequisite for existing methods to achieve counterfactual fairness is the prior human knowledge of the causal model for the data.
In this work, we address the problem of counterfactually fair prediction from observational data without given causal models by proposing a novel framework CLAIRE.
arXiv Detail & Related papers (2023-07-17T04:08:29Z) - Counterfactual Fair Opportunity: Measuring Decision Model Fairness with
Counterfactual Reasoning [5.626570248105078]
This work aims to unveil unfair model behaviors using counterfactual reasoning in the case of fairness under unawareness setting.
A counterfactual version of equal opportunity named counterfactual fair opportunity is defined and two novel metrics that analyze the sensitive information of counterfactual samples are introduced.
arXiv Detail & Related papers (2023-02-16T09:13:53Z) - Fair Enough: Standardizing Evaluation and Model Selection for Fairness
Research in NLP [64.45845091719002]
Modern NLP systems exhibit a range of biases, which a growing literature on model debiasing attempts to correct.
This paper seeks to clarify the current situation and plot a course for meaningful progress in fair learning.
arXiv Detail & Related papers (2023-02-11T14:54:00Z) - A Sandbox Tool to Bias(Stress)-Test Fairness Algorithms [19.86635585740634]
We present the conceptual idea and a first implementation of a bias-injection sandbox tool to investigate fairness consequences of various biases.
Unlike existing toolkits, ours provides a controlled environment to counterfactually inject biases in the ML pipeline.
In particular, we can test whether a given remedy can alleviate the injected bias by comparing the predictions resulting after the intervention with true labels in the unbiased regime-that is, before any bias injection.
arXiv Detail & Related papers (2022-04-21T16:12:19Z) - Towards Equal Opportunity Fairness through Adversarial Learning [64.45845091719002]
Adversarial training is a common approach for bias mitigation in natural language processing.
We propose an augmented discriminator for adversarial training, which takes the target class as input to create richer features.
arXiv Detail & Related papers (2022-03-12T02:22:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.