Integrated Decision Gradients: Compute Your Attributions Where the Model
Makes Its Decision
- URL: http://arxiv.org/abs/2305.20052v2
- Date: Mon, 18 Dec 2023 18:05:26 GMT
- Title: Integrated Decision Gradients: Compute Your Attributions Where the Model
Makes Its Decision
- Authors: Chase Walker, Sumit Jha, Kenny Chen, Rickard Ewetz
- Abstract summary: We propose an attribution algorithm called integrated decision gradients (IDG)
IDG focuses on integrating gradients from the region of the path where the model makes its decision, i.e., the portion of the path where the output logit rapidly transitions from zero to its final value.
We minimize the errors within the sum approximation of the path integral by utilizing non-uniform subdivisions determined by adaptive sampling.
- Score: 9.385886214196479
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Attribution algorithms are frequently employed to explain the decisions of
neural network models. Integrated Gradients (IG) is an influential attribution
method due to its strong axiomatic foundation. The algorithm is based on
integrating the gradients along a path from a reference image to the input
image. Unfortunately, it can be observed that gradients computed from regions
where the output logit changes minimally along the path provide poor
explanations for the model decision, which is called the saturation effect
problem. In this paper, we propose an attribution algorithm called integrated
decision gradients (IDG). The algorithm focuses on integrating gradients from
the region of the path where the model makes its decision, i.e., the portion of
the path where the output logit rapidly transitions from zero to its final
value. This is practically realized by scaling each gradient by the derivative
of the output logit with respect to the path. The algorithm thereby provides a
principled solution to the saturation problem. Additionally, we minimize the
errors within the Riemann sum approximation of the path integral by utilizing
non-uniform subdivisions determined by adaptive sampling. In the evaluation on
ImageNet, it is demonstrated that IDG outperforms IG, Left-IG, Guided IG, and
adversarial gradient integration both qualitatively and quantitatively using
standard insertion and deletion metrics across three common models.
Related papers
- Deep Manifold Learning with Graph Mining [80.84145791017968]
We propose a novel graph deep model with a non-gradient decision layer for graph mining.
The proposed model has achieved state-of-the-art performance compared to the current models.
arXiv Detail & Related papers (2022-07-18T04:34:08Z) - Geometrically Guided Integrated Gradients [0.3867363075280543]
We introduce an interpretability method called "geometrically-guided integrated gradients"
Our method explores the model's dynamic behavior from multiple scaled versions of the input and captures the best possible attribution for each input.
We also propose a "model perturbation" sanity check to complement the traditionally used "model randomization" test.
arXiv Detail & Related papers (2022-06-13T05:05:43Z) - Cogradient Descent for Dependable Learning [64.02052988844301]
We propose a dependable learning based on Cogradient Descent (CoGD) algorithm to address the bilinear optimization problem.
CoGD is introduced to solve bilinear problems when one variable is with sparsity constraint.
It can also be used to decompose the association of features and weights, which further generalizes our method to better train convolutional neural networks (CNNs)
arXiv Detail & Related papers (2021-06-20T04:28:20Z) - Guided Integrated Gradients: An Adaptive Path Method for Removing Noise [9.792727625917083]
Integrated Gradients (IG) is a commonly used feature attribution method for deep neural networks.
We show that one of the causes of the problem is the accumulation of noise along the IG path.
We propose adapting the attribution path itself -- conditioning the path not just on the image but also on the model being explained.
arXiv Detail & Related papers (2021-06-17T20:00:55Z) - Integrated Grad-CAM: Sensitivity-Aware Visual Explanation of Deep
Convolutional Networks via Integrated Gradient-Based Scoring [26.434705114982584]
Grad-CAM is a popular solution that provides such a visualization by combining the activation maps obtained from the model.
We introduce a solution to tackle this problem by computing the path integral of the gradient-based terms in Grad-CAM.
We conduct a thorough analysis to demonstrate the improvement achieved by our method in measuring the importance of the extracted representations for the CNN's predictions.
arXiv Detail & Related papers (2021-02-15T19:21:46Z) - Investigating Saturation Effects in Integrated Gradients [5.366801257602863]
We propose a variant of IntegratedGradients which primarily captures gradients in unsaturated regions.
We find that this attribution technique shows higher model faithfulness and lower sensitivity to noise com-pared with standard Integrated Gradients.
arXiv Detail & Related papers (2020-10-23T22:48:02Z) - Deep Variational Network Toward Blind Image Restoration [60.45350399661175]
Blind image restoration is a common yet challenging problem in computer vision.
We propose a novel blind image restoration method, aiming to integrate both the advantages of them.
Experiments on two typical blind IR tasks, namely image denoising and super-resolution, demonstrate that the proposed method achieves superior performance over current state-of-the-arts.
arXiv Detail & Related papers (2020-08-25T03:30:53Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z) - Understanding Integrated Gradients with SmoothTaylor for Deep Neural
Network Attribution [70.78655569298923]
Integrated Gradients as an attribution method for deep neural network models offers simple implementability.
It suffers from noisiness of explanations which affects the ease of interpretability.
The SmoothGrad technique is proposed to solve the noisiness issue and smoothen the attribution maps of any gradient-based attribution method.
arXiv Detail & Related papers (2020-04-22T10:43:19Z) - Towards Better Understanding of Adaptive Gradient Algorithms in
Generative Adversarial Nets [71.05306664267832]
Adaptive algorithms perform gradient updates using the history of gradients and are ubiquitous in training deep neural networks.
In this paper we analyze a variant of OptimisticOA algorithm for nonconcave minmax problems.
Our experiments show that adaptive GAN non-adaptive gradient algorithms can be observed empirically.
arXiv Detail & Related papers (2019-12-26T22:10:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.