Improving Explainability of Disentangled Representations using
Multipath-Attribution Mappings
- URL: http://arxiv.org/abs/2306.09035v1
- Date: Thu, 15 Jun 2023 10:52:29 GMT
- Title: Improving Explainability of Disentangled Representations using
Multipath-Attribution Mappings
- Authors: Lukas Klein, Jo\~ao B. S. Carvalho, Mennatallah El-Assady, Paolo
Penna, Joachim M. Buhmann, Paul F. Jaeger
- Abstract summary: We propose a framework that utilizes interpretable disentangled representations for downstream-task prediction.
We demonstrate the effectiveness of our approach on a synthetic benchmark suite and two medical datasets.
- Score: 12.145748796751619
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Explainable AI aims to render model behavior understandable by humans, which
can be seen as an intermediate step in extracting causal relations from
correlative patterns. Due to the high risk of possible fatal decisions in
image-based clinical diagnostics, it is necessary to integrate explainable AI
into these safety-critical systems. Current explanatory methods typically
assign attribution scores to pixel regions in the input image, indicating their
importance for a model's decision. However, they fall short when explaining why
a visual feature is used. We propose a framework that utilizes interpretable
disentangled representations for downstream-task prediction. Through
visualizing the disentangled representations, we enable experts to investigate
possible causation effects by leveraging their domain knowledge. Additionally,
we deploy a multi-path attribution mapping for enriching and validating
explanations. We demonstrate the effectiveness of our approach on a synthetic
benchmark suite and two medical datasets. We show that the framework not only
acts as a catalyst for causal relation extraction but also enhances model
robustness by enabling shortcut detection without the need for testing under
distribution shifts.
Related papers
- Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data [14.991686165405959]
We introduce a semi-automated framework for the identification of spurious behavior from both data and model perspective.
This allows the retrieval of spurious data points and the detection of model circuits that encode the associated prediction rules.
We show the applicability of our framework using four medical datasets, featuring controlled and real-world spurious correlations.
arXiv Detail & Related papers (2025-01-23T16:39:09Z) - Unsupervised Model Diagnosis [49.36194740479798]
This paper proposes Unsupervised Model Diagnosis (UMO) to produce semantic counterfactual explanations without any user guidance.
Our approach identifies and visualizes changes in semantics, and then matches these changes to attributes from wide-ranging text sources.
arXiv Detail & Related papers (2024-10-08T17:59:03Z) - Identifiable Latent Neural Causal Models [82.14087963690561]
Causal representation learning seeks to uncover latent, high-level causal representations from low-level observed data.
We determine the types of distribution shifts that do contribute to the identifiability of causal representations.
We translate our findings into a practical algorithm, allowing for the acquisition of reliable latent causal representations.
arXiv Detail & Related papers (2024-03-23T04:13:55Z) - Diffusion Model with Cross Attention as an Inductive Bias for Disentanglement [58.9768112704998]
Disentangled representation learning strives to extract the intrinsic factors within observed data.
We introduce a new perspective and framework, demonstrating that diffusion models with cross-attention can serve as a powerful inductive bias.
This is the first work to reveal the potent disentanglement capability of diffusion models with cross-attention, requiring no complex designs.
arXiv Detail & Related papers (2024-02-15T05:07:54Z) - Detecting Spurious Correlations via Robust Visual Concepts in Real and
AI-Generated Image Classification [12.992095539058022]
We introduce a general-purpose method that efficiently detects potential spurious correlations.
The proposed method provides intuitive explanations while eliminating the need for pixel-level annotations.
Our method is also suitable for detecting spurious correlations that may propagate to downstream applications originating from generative models.
arXiv Detail & Related papers (2023-11-03T01:12:35Z) - Causality-Driven One-Shot Learning for Prostate Cancer Grading from MRI [1.049712834719005]
We present a novel method to automatically classify medical images that learns and leverages weak causal signals in the image.
Our framework consists of a convolutional neural network backbone and a causality-extractor module.
Our findings show that causal relationships among features play a crucial role in enhancing the model's ability to discern relevant information.
arXiv Detail & Related papers (2023-09-19T16:08:33Z) - Causal Transportability for Visual Recognition [70.13627281087325]
We show that standard classifiers fail because the association between images and labels is not transportable across settings.
We then show that the causal effect, which severs all sources of confounding, remains invariant across domains.
This motivates us to develop an algorithm to estimate the causal effect for image classification.
arXiv Detail & Related papers (2022-04-26T15:02:11Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Proactive Pseudo-Intervention: Causally Informed Contrastive Learning
For Interpretable Vision Models [103.64435911083432]
We present a novel contrastive learning strategy called it Proactive Pseudo-Intervention (PPI)
PPI leverages proactive interventions to guard against image features with no causal relevance.
We also devise a novel causally informed salience mapping module to identify key image pixels to intervene, and show it greatly facilitates model interpretability.
arXiv Detail & Related papers (2020-12-06T20:30:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.