A fuzzy-rough uncertainty measure to discover bias encoded explicitly or
implicitly in features of structured pattern classification datasets
- URL: http://arxiv.org/abs/2108.09098v1
- Date: Fri, 20 Aug 2021 10:27:32 GMT
- Title: A fuzzy-rough uncertainty measure to discover bias encoded explicitly or
implicitly in features of structured pattern classification datasets
- Authors: Gonzalo N\'apoles, Lisa Koutsoviti Koumeri
- Abstract summary: We study the existence of bias encoded implicitly in non-protected features as defined by the correlation between protected and unprotected attributes.
We conduct a sensitivity analysis to determine the fuzzy operatorsand distance function that best capture change in the boundary regions.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The need to measure bias encoded in tabular data that are used to solve
pattern recognition problems is widely recognized by academia, legislators and
enterprises alike. In previous work, we proposed a bias quantification measure,
called fuzzy-rough uncer-tainty, which relies on the fuzzy-rough set theory.
The intuition dictates that protected features should not change the
fuzzy-rough boundary regions of a decision class significantly. The extent to
which this happens is a proxy for bias expressed as uncertainty in
adecision-making context. Our measure's main advantage is that it does not
depend on any machine learning prediction model but adistance function. In this
paper, we extend our study by exploring the existence of bias encoded
implicitly in non-protected featuresas defined by the correlation between
protected and unprotected attributes. This analysis leads to four scenarios
that domain experts should evaluate before deciding how to tackle bias. In
addition, we conduct a sensitivity analysis to determine the fuzzy operatorsand
distance function that best capture change in the boundary regions.
Related papers
- Unlearning-based Neural Interpretations [51.99182464831169]
We show that current baselines defined using static functions are biased, fragile and manipulable.
We propose UNI to compute an (un)learnable, debiased and adaptive baseline by perturbing the input towards an unlearning direction of steepest ascent.
arXiv Detail & Related papers (2024-10-10T16:02:39Z) - Detecting and Identifying Selection Structure in Sequential Data [53.24493902162797]
We argue that the selective inclusion of data points based on latent objectives is common in practical situations, such as music sequences.
We show that selection structure is identifiable without any parametric assumptions or interventional experiments.
We also propose a provably correct algorithm to detect and identify selection structures as well as other types of dependencies.
arXiv Detail & Related papers (2024-06-29T20:56:34Z) - Debiasing Machine Learning Models by Using Weakly Supervised Learning [3.3298048942057523]
We tackle the problem of bias mitigation of algorithmic decisions in a setting where both the output of the algorithm and the sensitive variable are continuous.
Typical examples are unfair decisions made with respect to the age or the financial status.
Our bias mitigation strategy is a weakly supervised learning method which requires that a small portion of the data can be measured in a fair manner.
arXiv Detail & Related papers (2024-02-23T18:11:32Z) - Towards Assumption-free Bias Mitigation [47.5131072745805]
We propose an assumption-free framework to detect the related attributes automatically by modeling feature interaction for bias mitigation.
Experimental results on four real-world datasets demonstrate that our proposed framework can significantly alleviate unfair prediction behaviors.
arXiv Detail & Related papers (2023-07-09T05:55:25Z) - Measuring Implicit Bias Using SHAP Feature Importance and Fuzzy
Cognitive Maps [1.9739269019020032]
In this paper, we integrate the concepts of feature importance with implicit bias in the context of pattern classification.
The amount of bias towards protected features might differ depending on whether the features are numerically or categorically encoded.
arXiv Detail & Related papers (2023-05-16T12:31:36Z) - Counterfactual Reasoning for Bias Evaluation and Detection in a Fairness
under Unawareness setting [6.004889078682389]
Current AI regulations require discarding sensitive features in the algorithm's decision-making process to prevent unfair outcomes.
We propose a way to reveal the potential hidden bias of a machine learning model that can persist even when sensitive features are discarded.
arXiv Detail & Related papers (2023-02-16T10:36:18Z) - Uncertainty Estimates of Predictions via a General Bias-Variance
Decomposition [7.811916700683125]
We introduce a bias-variance decomposition for proper scores, giving rise to the Bregman Information as the variance term.
We showcase the practical relevance of this decomposition on several downstream tasks, including model ensembles and confidence regions.
arXiv Detail & Related papers (2022-10-21T21:24:37Z) - Data-Driven Influence Functions for Optimization-Based Causal Inference [105.5385525290466]
We study a constructive algorithm that approximates Gateaux derivatives for statistical functionals by finite differencing.
We study the case where probability distributions are not known a priori but need to be estimated from data.
arXiv Detail & Related papers (2022-08-29T16:16:22Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - CertainNet: Sampling-free Uncertainty Estimation for Object Detection [65.28989536741658]
Estimating the uncertainty of a neural network plays a fundamental role in safety-critical settings.
In this work, we propose a novel sampling-free uncertainty estimation method for object detection.
We call it CertainNet, and it is the first to provide separate uncertainties for each output signal: objectness, class, location and size.
arXiv Detail & Related papers (2021-10-04T17:59:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.