Navigating Explanatory Multiverse Through Counterfactual Path Geometry
- URL: http://arxiv.org/abs/2306.02786v3
- Date: Fri, 3 May 2024 18:42:14 GMT
- Title: Navigating Explanatory Multiverse Through Counterfactual Path Geometry
- Authors: Kacper Sokol, Edward Small, Yueqing Xuan,
- Abstract summary: We introduce the novel concept of explanatory multiverse.
We show how to navigate, reason about and compare the geometry of these trajectories.
We propose an all-in-one metric, called opportunity potential, to quantify them.
- Score: 5.109188339767978
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Counterfactual explanations are the de facto standard when tasked with interpreting decisions of (opaque) predictive models. Their generation is often subject to algorithmic and domain-specific constraints -- such as density-based feasibility, and attribute (im)mutability or directionality of change -- that aim to maximise their real-life utility. In addition to desiderata with respect to the counterfactual instance itself, existence of a viable path connecting it with the factual data point, known as algorithmic recourse, has become an important technical consideration. While both of these requirements ensure that the steps of the journey as well as its destination are admissible, current literature neglects the multiplicity of such counterfactual paths. To address this shortcoming we introduce the novel concept of explanatory multiverse that encompasses all the possible counterfactual journeys. We then show how to navigate, reason about and compare the geometry of these trajectories with two methods: vector spaces and graphs. To this end, we overview their spacial properties -- such as affinity, branching, divergence and possible future convergence -- and propose an all-in-one metric, called opportunity potential, to quantify them. Implementing this (possibly interactive) explanatory process grants explainees agency by allowing them to select counterfactuals based on the properties of the journey leading to them in addition to their absolute differences. We show the flexibility, benefit and efficacy of such an approach through examples and quantitative evaluation on the German Credit and MNIST data sets.
Related papers
- Rethinking Distance Metrics for Counterfactual Explainability [53.436414009687]
We investigate a framing for counterfactual generation methods that considers counterfactuals not as independent draws from a region around the reference, but as jointly sampled with the reference from the underlying data distribution.
We derive a distance metric, tailored for counterfactual similarity that can be applied to a broad range of settings.
arXiv Detail & Related papers (2024-10-18T15:06:50Z) - LLM Uncertainty Quantification through Directional Entailment Graph and Claim Level Response Augmentation [5.255129053741665]
Large language models (LLMs) have showcased superior capabilities in sophisticated tasks across various domains, stemming from basic question-answer (QA)
This paper presents a novel way to evaluate the uncertainty that captures the directional instability, by constructing a directional graph from entailment probabilities.
We also provide a way to incorporate the existing work's semantics uncertainty with our proposed layer.
arXiv Detail & Related papers (2024-07-01T06:11:30Z) - Enhancing Counterfactual Explanation Search with Diffusion Distance and Directional Coherence [0.0]
A pressing issue in the adoption of AI models is the increasing demand for more human-centric explanations of their predictions.
We propose and test the incorporation of two novel biases to enhance the search for effective counterfactual explanations.
arXiv Detail & Related papers (2024-04-19T11:47:17Z) - Nonparametric Partial Disentanglement via Mechanism Sparsity: Sparse
Actions, Interventions and Sparse Temporal Dependencies [58.179981892921056]
This work introduces a novel principle for disentanglement we call mechanism sparsity regularization.
We propose a representation learning method that induces disentanglement by simultaneously learning the latent factors.
We show that the latent factors can be recovered by regularizing the learned causal graph to be sparse.
arXiv Detail & Related papers (2024-01-10T02:38:21Z) - Enriching Disentanglement: From Logical Definitions to Quantitative Metrics [59.12308034729482]
Disentangling the explanatory factors in complex data is a promising approach for data-efficient representation learning.
We establish relationships between logical definitions and quantitative metrics to derive theoretically grounded disentanglement metrics.
We empirically demonstrate the effectiveness of the proposed metrics by isolating different aspects of disentangled representations.
arXiv Detail & Related papers (2023-05-19T08:22:23Z) - Latent Traversals in Generative Models as Potential Flows [113.4232528843775]
We propose to model latent structures with a learned dynamic potential landscape.
Inspired by physics, optimal transport, and neuroscience, these potential landscapes are learned as physically realistic partial differential equations.
Our method achieves both more qualitatively and quantitatively disentangled trajectories than state-of-the-art baselines.
arXiv Detail & Related papers (2023-04-25T15:53:45Z) - Implicit Bayes Adaptation: A Collaborative Transport Approach [25.96406219707398]
We show that domain adaptation is rooted in the intrinsic representations of the respective data, which are inherently lying in a non-linear submanifold embedded in a higher dimensional Euclidean space.
We show that this is tantamount to an implicit Bayesian framework, which we demonstrate to be viable for a more robust and better-performing approach to domain adaptation.
arXiv Detail & Related papers (2023-04-17T14:13:40Z) - Simplified Continuous High Dimensional Belief Space Planning with
Adaptive Probabilistic Belief-dependent Constraints [9.061408029414453]
Under uncertainty in partially observable domains, also known as Belief Space Planning, online decision making is a fundamental problem.
We present a technique to adaptively accept or discard a candidate action sequence with respect to a probabilistic belief-dependent constraint.
We apply our method to active SLAM, a highly challenging problem of high dimensional Belief Space Planning.
arXiv Detail & Related papers (2023-02-13T21:22:47Z) - Towards Explainable Land Cover Mapping: a Counterfactual-based Strategy [9.180712157534606]
We propose a generative adversarial counterfactual approach for satellite image time series in a multi-class setting for the land cover classification task.
One of the distinctive features of the proposed approach is the lack of prior assumption on the targeted class for a given counterfactual explanation.
arXiv Detail & Related papers (2023-01-04T10:17:16Z) - Exploring the Trade-off between Plausibility, Change Intensity and
Adversarial Power in Counterfactual Explanations using Multi-objective
Optimization [73.89239820192894]
We argue that automated counterfactual generation should regard several aspects of the produced adversarial instances.
We present a novel framework for the generation of counterfactual examples.
arXiv Detail & Related papers (2022-05-20T15:02:53Z) - Learning Conditional Invariance through Cycle Consistency [60.85059977904014]
We propose a novel approach to identify meaningful and independent factors of variation in a dataset.
Our method involves two separate latent subspaces for the target property and the remaining input information.
We demonstrate on synthetic and molecular data that our approach identifies more meaningful factors which lead to sparser and more interpretable models.
arXiv Detail & Related papers (2021-11-25T17:33:12Z) - You Mostly Walk Alone: Analyzing Feature Attribution in Trajectory
Prediction [52.442129609979794]
Recent deep learning approaches for trajectory prediction show promising performance.
It remains unclear which features such black-box models actually learn to use for making predictions.
This paper proposes a procedure that quantifies the contributions of different cues to model performance.
arXiv Detail & Related papers (2021-10-11T14:24:15Z) - Learning Disentangled Representations with Latent Variation
Predictability [102.4163768995288]
This paper defines the variation predictability of latent disentangled representations.
Within an adversarial generation process, we encourage variation predictability by maximizing the mutual information between latent variations and corresponding image pairs.
We develop an evaluation metric that does not rely on the ground-truth generative factors to measure the disentanglement of latent representations.
arXiv Detail & Related papers (2020-07-25T08:54:26Z) - Orientation Attentive Robotic Grasp Synthesis with Augmented Grasp Map
Representation [62.79160608266713]
morphological characteristics in objects may offer a wide range of plausible grasping orientations that obfuscates the visual learning of robotic grasping.
Existing grasp generation approaches are cursed to construct discontinuous grasp maps by aggregating annotations for drastically different orientations per grasping point.
We propose a novel augmented grasp map representation, suitable for pixel-wise synthesis, that locally disentangles grasping orientations by partitioning the angle space into multiple bins.
arXiv Detail & Related papers (2020-06-09T08:54:54Z) - Explainable Link Prediction for Emerging Entities in Knowledge Graphs [44.87285668747474]
Cross-domain knowledge graphs suffer from inherent incompleteness and sparsity.
Link prediction can alleviate this by inferring a target entity, given a source entity and a query relation.
We propose an inductive representation learning framework that is able to learn representations of previously unseen entities.
arXiv Detail & Related papers (2020-05-01T22:17:37Z) - Scenario-Transferable Semantic Graph Reasoning for Interaction-Aware
Probabilistic Prediction [29.623692599892365]
Accurately predicting the possible behaviors of traffic participants is an essential capability for autonomous vehicles.
We propose a novel generic representation for various driving environments by taking the advantage of semantics and domain knowledge.
arXiv Detail & Related papers (2020-04-07T00:34:36Z) - A Bounded Measure for Estimating the Benefit of Visualization [3.8360246117087473]
Information theory can be used to analyze the cost-benefit of visualization processes.
The current measure of benefit contains an unbounded term that is neither easy to estimate nor intuitive to interpret.
We propose to revise the existing cost-benefit measure by replacing the unbounded term with a bounded one.
arXiv Detail & Related papers (2020-02-12T23:39:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.