SAFE: Saliency-Aware Counterfactual Explanations for DNN-based Automated
Driving Systems
- URL: http://arxiv.org/abs/2307.15786v1
- Date: Fri, 28 Jul 2023 19:56:01 GMT
- Title: SAFE: Saliency-Aware Counterfactual Explanations for DNN-based Automated
Driving Systems
- Authors: Amir Samadi, Amir Shirian, Konstantinos Koufos, Kurt Debattista and
Mehrdad Dianati
- Abstract summary: A CF explainer computes the minimum modifications required to cross the model's decision boundary.
Current deep generative CF models often work with user-selected features rather than focusing on the discriminative features of the black-box model.
We propose in this paper a novel approach that leverages saliency maps to generate more informative CF explanations.
- Score: 10.40211479079817
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A CF explainer identifies the minimum modifications in the input that would
alter the model's output to its complement. In other words, a CF explainer
computes the minimum modifications required to cross the model's decision
boundary. Current deep generative CF models often work with user-selected
features rather than focusing on the discriminative features of the black-box
model. Consequently, such CF examples may not necessarily lie near the decision
boundary, thereby contradicting the definition of CFs. To address this issue,
we propose in this paper a novel approach that leverages saliency maps to
generate more informative CF explanations. Source codes are available at:
https://github.com/Amir-Samadi//Saliency_Aware_CF.
Related papers
- The Gaussian Discriminant Variational Autoencoder (GdVAE): A Self-Explainable Model with Counterfactual Explanations [6.741417019751439]
Visual counterfactual explanation (CF) methods modify image concepts to change a prediction to a predefined outcome.
We introduce the GdVAE, a self-explainable model based on a conditional variational autoencoder (CVAE) and integrated CF explanations.
The consistency of CFs is improved by regularizing the latent space with the explainer function.
arXiv Detail & Related papers (2024-09-19T17:58:15Z) - Revisiting Catastrophic Forgetting in Large Language Model Tuning [79.70722658190097]
Catastrophic Forgetting (CF) means models forgetting previously acquired knowledge when learning new data.
This paper takes the first step to reveal the direct link between the flatness of the model loss landscape and the extent of CF in the field of large language models.
Experiments on three widely-used fine-tuning datasets, spanning different model scales, demonstrate the effectiveness of our method in alleviating CF.
arXiv Detail & Related papers (2024-06-07T11:09:13Z) - Diffusion-TTA: Test-time Adaptation of Discriminative Models via
Generative Feedback [97.0874638345205]
generative models can be great test-time adapters for discriminative models.
Our method, Diffusion-TTA, adapts pre-trained discriminative models to each unlabelled example in the test set.
We show Diffusion-TTA significantly enhances the accuracy of various large-scale pre-trained discriminative models.
arXiv Detail & Related papers (2023-11-27T18:59:53Z) - Diffusion models for probabilistic programming [56.47577824219207]
Diffusion Model Variational Inference (DMVI) is a novel method for automated approximate inference in probabilistic programming languages (PPLs)
DMVI is easy to implement, allows hassle-free inference in PPLs without the drawbacks of, e.g., variational inference using normalizing flows, and does not make any constraints on the underlying neural network model.
arXiv Detail & Related papers (2023-11-01T12:17:05Z) - Co(ve)rtex: ML Models as storage channels and their (mis-)applications [2.792027541710663]
In machine learning systems, don't-care states and undefined behavior have been shown to be sources of significant vulnerabilities.
We consider the ML model as a storage channel with a capacity that increases with over parameterization.
We develop optimizations to improve the capacity in this case, including a novel ML-specific substitution based error correction protocol.
arXiv Detail & Related papers (2023-07-17T19:57:10Z) - DualCF: Efficient Model Extraction Attack from Counterfactual
Explanations [57.46134660974256]
Cloud service providers have launched Machine-Learning-as-a-Service platforms to allow users to access large-scale cloudbased models via APIs.
Such extra information inevitably causes the cloud models to be more vulnerable to extraction attacks.
We propose a novel simple yet efficient querying strategy to greatly enhance the querying efficiency to steal a classification model.
arXiv Detail & Related papers (2022-05-13T08:24:43Z) - MGDCF: Distance Learning via Markov Graph Diffusion for Neural
Collaborative Filtering [96.65234340724237]
We show the equivalence between some state-of-the-art GNN-based CF models and a traditional 1-layer NRL model based on context encoding.
We present Markov Graph Diffusion Collaborative Filtering (MGDCF) to generalize some state-of-the-art GNN-based CF models.
arXiv Detail & Related papers (2022-04-05T17:24:32Z) - ReLACE: Reinforcement Learning Agent for Counterfactual Explanations of
Arbitrary Predictive Models [6.939617874336667]
We introduce a model-agnostic algorithm to generate optimal counterfactual explanations.
Our method is easily applied to any black-box model, as this resembles the environment that the DRL agent interacts with.
In addition, we develop an algorithm to extract explainable decision rules from the DRL agent's policy, so as to make the process of generating CFs itself transparent.
arXiv Detail & Related papers (2021-10-22T17:08:49Z) - CounterNet: End-to-End Training of Prediction Aware Counterfactual
Explanations [12.313007847721215]
CounterNet is an end-to-end learning framework which integrates predictive model training and the generation of counterfactual (CF) explanations.
Unlike post-hoc methods, CounterNet enables the optimization of the CF explanation generation only once together with the predictive model.
Our experiments on multiple real-world datasets show that CounterNet generates high-quality predictions.
arXiv Detail & Related papers (2021-09-15T20:09:13Z) - An exact counterfactual-example-based approach to tree-ensemble models
interpretability [0.0]
High-performance models do not exhibit the necessary transparency to make their decisions fully understandable.
We could derive an exact geometrical characterisation of their decision regions under the form of a collection of multidimensional intervals.
An adaptation to reasoning on regression problems is also envisaged.
arXiv Detail & Related papers (2021-05-31T09:32:46Z) - Explanation-Guided Training for Cross-Domain Few-Shot Classification [96.12873073444091]
Cross-domain few-shot classification task (CD-FSC) combines few-shot classification with the requirement to generalize across domains represented by datasets.
We introduce a novel training approach for existing FSC models.
We show that explanation-guided training effectively improves the model generalization.
arXiv Detail & Related papers (2020-07-17T07:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.