Don't Lie to Me! Robust and Efficient Explainability with Verified
Perturbation Analysis
- URL: http://arxiv.org/abs/2202.07728v2
- Date: Sat, 18 Mar 2023 10:45:10 GMT
- Title: Don't Lie to Me! Robust and Efficient Explainability with Verified
Perturbation Analysis
- Authors: Thomas Fel, Melanie Ducoffe, David Vigouroux, Remi Cadene, Mikael
Capelle, Claire Nicodeme, Thomas Serre
- Abstract summary: We introduce EVA -- the first explainability method guarantee to have an exhaustive exploration of a perturbation space.
We leverage the beneficial properties of verified perturbation analysis to efficiently characterize the input variables that are most likely to drive the model decision.
- Score: 6.15738282053772
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A variety of methods have been proposed to try to explain how deep neural
networks make their decisions. Key to those approaches is the need to sample
the pixel space efficiently in order to derive importance maps. However, it has
been shown that the sampling methods used to date introduce biases and other
artifacts, leading to inaccurate estimates of the importance of individual
pixels and severely limit the reliability of current explainability methods.
Unfortunately, the alternative -- to exhaustively sample the image space is
computationally prohibitive. In this paper, we introduce EVA (Explaining using
Verified perturbation Analysis) -- the first explainability method guarantee to
have an exhaustive exploration of a perturbation space. Specifically, we
leverage the beneficial properties of verified perturbation analysis -- time
efficiency, tractability and guaranteed complete coverage of a manifold -- to
efficiently characterize the input variables that are most likely to drive the
model decision. We evaluate the approach systematically and demonstrate
state-of-the-art results on multiple benchmarks.
Related papers
- Towards stable real-world equation discovery with assessing
differentiating quality influence [52.2980614912553]
We propose alternatives to the commonly used finite differences-based method.
We evaluate these methods in terms of applicability to problems, similar to the real ones, and their ability to ensure the convergence of equation discovery algorithms.
arXiv Detail & Related papers (2023-11-09T23:32:06Z) - Observation-Guided Diffusion Probabilistic Models [41.749374023639156]
We propose a novel diffusion-based image generation method called the observation-guided diffusion probabilistic model (OGDM)
Our approach reestablishes the training objective by integrating the guidance of the observation process with the Markov chain.
We demonstrate the effectiveness of our training algorithm using diverse inference techniques on strong diffusion model baselines.
arXiv Detail & Related papers (2023-10-06T06:29:06Z) - Bayesian Hierarchical Models for Counterfactual Estimation [12.159830463756341]
We propose a probabilistic paradigm to estimate a diverse set of counterfactuals.
We treat the perturbations as random variables endowed with prior distribution functions.
A gradient based sampler with superior convergence characteristics efficiently computes the posterior samples.
arXiv Detail & Related papers (2023-01-21T00:21:11Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Information-Theoretic Odometry Learning [83.36195426897768]
We propose a unified information theoretic framework for learning-motivated methods aimed at odometry estimation.
The proposed framework provides an elegant tool for performance evaluation and understanding in information-theoretic language.
arXiv Detail & Related papers (2022-03-11T02:37:35Z) - Fidelity of Interpretability Methods and Perturbation Artifacts in
Neural Networks [5.439020425819001]
Post-hoc interpretability methods aim to quantify the importance of input features with respect to the class probabilities.
A popular approach to evaluate interpretability methods is to perturb input features deemed important for a given prediction and observe the decrease in accuracy.
We propose a method for estimating the impact of such artifacts on the fidelity estimation by utilizing model accuracy curves from perturbing input features.
arXiv Detail & Related papers (2022-03-06T10:14:09Z) - Time to Focus: A Comprehensive Benchmark Using Time Series Attribution
Methods [4.9449660544238085]
The paper focuses on time series analysis and benchmark several state-of-the-art attribution methods.
The presented experiments involve gradient-based and perturbation-based attribution methods.
The findings accentuate that choosing the best-suited attribution method is strongly correlated with the desired use case.
arXiv Detail & Related papers (2022-02-08T10:06:13Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Learning Conditional Invariance through Cycle Consistency [60.85059977904014]
We propose a novel approach to identify meaningful and independent factors of variation in a dataset.
Our method involves two separate latent subspaces for the target property and the remaining input information.
We demonstrate on synthetic and molecular data that our approach identifies more meaningful factors which lead to sparser and more interpretable models.
arXiv Detail & Related papers (2021-11-25T17:33:12Z) - DANCE: Enhancing saliency maps using decoys [35.46266461621123]
We propose a framework that improves the robustness of saliency methods by following a two-step procedure.
First, we introduce a perturbation mechanism that subtly varies the input sample without changing its intermediate representations.
Second, we compute saliency maps for perturbed samples and propose a new method to aggregate saliency maps.
arXiv Detail & Related papers (2020-02-03T01:21:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.