A New Paradigm for Counterfactual Reasoning in Fairness and Recourse
- URL: http://arxiv.org/abs/2401.13935v1
- Date: Thu, 25 Jan 2024 04:28:39 GMT
- Title: A New Paradigm for Counterfactual Reasoning in Fairness and Recourse
- Authors: Lucius E.J. Bynum, Joshua R. Loftus, Julia Stoyanovich
- Abstract summary: The traditional paradigm for counterfactual reasoning in this literature is the interventional counterfactual.
An inherent limitation of this paradigm is that some demographic interventions may not translate into the formalisms of interventional counterfactuals.
In this work, we explore a new paradigm based instead on the backtracking counterfactual.
- Score: 12.119272303766056
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Counterfactuals and counterfactual reasoning underpin numerous techniques for
auditing and understanding artificial intelligence (AI) systems. The
traditional paradigm for counterfactual reasoning in this literature is the
interventional counterfactual, where hypothetical interventions are imagined
and simulated. For this reason, the starting point for causal reasoning about
legal protections and demographic data in AI is an imagined intervention on a
legally-protected characteristic, such as ethnicity, race, gender, disability,
age, etc. We ask, for example, what would have happened had your race been
different? An inherent limitation of this paradigm is that some demographic
interventions -- like interventions on race -- may not translate into the
formalisms of interventional counterfactuals. In this work, we explore a new
paradigm based instead on the backtracking counterfactual, where rather than
imagine hypothetical interventions on legally-protected characteristics, we
imagine alternate initial conditions while holding these characteristics fixed.
We ask instead, what would explain a counterfactual outcome for you as you
actually are or could be? This alternate framework allows us to address many of
the same social concerns, but to do so while asking fundamentally different
questions that do not rely on demographic interventions.
Related papers
- Longitudinal Counterfactuals: Constraints and Opportunities [59.11233767208572]
We propose using longitudinal data to assess and improve plausibility in counterfactuals.
We develop a metric that compares longitudinal differences to counterfactual differences, allowing us to evaluate how similar a counterfactual is to prior observed changes.
arXiv Detail & Related papers (2024-02-29T20:17:08Z) - Explaining Text Classifiers with Counterfactual Representations [0.0]
We propose a simple method for generating counterfactuals by intervening in the space of text representations.
To validate our method, we conducted experiments first on a synthetic dataset and then on a realistic dataset of counterfactuals.
arXiv Detail & Related papers (2024-02-01T16:06:35Z) - Estimation of Counterfactual Interventions under Uncertainties [10.674015311238696]
"What should I have done differently to get the loan approved?"
"What should I have done differently to get the loan approved?"
"What should I have done differently to get the loan approved?"
arXiv Detail & Related papers (2023-09-15T11:41:23Z) - Emergent Causality and the Foundation of Consciousness [0.0]
We argue that in the absence of a $do$ operator, an intervention can be represented by a variable.
In a narrow sense this describes what it is to be aware, and is a mechanistic explanation of aspects of consciousness.
arXiv Detail & Related papers (2023-02-07T01:41:23Z) - Backtracking Counterfactuals [6.709991492637819]
We explore and formalise an alternative mode of counterfactual reasoning within the structural causal model (SCM) framework.
Despite ample evidence that humans backtrack, the present work constitutes to the best of our knowledge, the first general account and algorithmisation of backtracking counterfactuals.
arXiv Detail & Related papers (2022-11-01T14:07:30Z) - Neural Causal Models for Counterfactual Identification and Estimation [62.30444687707919]
We study the evaluation of counterfactual statements through neural models.
First, we show that neural causal models (NCMs) are expressive enough.
Second, we develop an algorithm for simultaneously identifying and estimating counterfactual distributions.
arXiv Detail & Related papers (2022-09-30T18:29:09Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - Empirical Estimates on Hand Manipulation are Recoverable: A Step Towards
Individualized and Explainable Robotic Support in Everyday Activities [80.37857025201036]
Key challenge for robotic systems is to figure out the behavior of another agent.
Processing correct inferences is especially challenging when (confounding) factors are not controlled experimentally.
We propose equipping robots with the necessary tools to conduct observational studies on people.
arXiv Detail & Related papers (2022-01-27T22:15:56Z) - Nested Counterfactual Identification from Arbitrary Surrogate
Experiments [95.48089725859298]
We study the identification of nested counterfactuals from an arbitrary combination of observations and experiments.
Specifically, we prove the counterfactual unnesting theorem (CUT), which allows one to map arbitrary nested counterfactuals to unnested ones.
arXiv Detail & Related papers (2021-07-07T12:51:04Z) - CausalCity: Complex Simulations with Agency for Causal Discovery and
Reasoning [68.74447489372037]
We present a high-fidelity simulation environment that is designed for developing algorithms for causal discovery and counterfactual reasoning.
A core component of our work is to introduce textitagency, such that it is simple to define and create complex scenarios.
We perform experiments with three state-of-the-art methods to create baselines and highlight the affordances of this environment.
arXiv Detail & Related papers (2021-06-25T00:21:41Z) - On Generating Plausible Counterfactual and Semi-Factual Explanations for
Deep Learning [15.965337956587373]
PlausIble Exceptionality-based Contrastive Explanations (PIECE), modifies all exceptional features in a test image to be normal from the perspective of the counterfactual class.
Two controlled experiments compare PIECE to others in the literature, showing that PIECE not only generates the most plausible counterfactuals on several measures, but also the best semifactuals.
arXiv Detail & Related papers (2020-09-10T14:48:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.