Investigating Saturation Effects in Integrated Gradients
- URL: http://arxiv.org/abs/2010.12697v1
- Date: Fri, 23 Oct 2020 22:48:02 GMT
- Title: Investigating Saturation Effects in Integrated Gradients
- Authors: Vivek Miglani and Narine Kokhlikyan and Bilal Alsallakh and Miguel
Martin and Orion Reblitz-Richardson
- Abstract summary: We propose a variant of IntegratedGradients which primarily captures gradients in unsaturated regions.
We find that this attribution technique shows higher model faithfulness and lower sensitivity to noise com-pared with standard Integrated Gradients.
- Score: 5.366801257602863
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Integrated Gradients has become a popular method for post-hoc model
interpretability. De-spite its popularity, the composition and relative impact
of different regions of the integral path are not well understood. We explore
these effects and find that gradients in saturated regions of this path, where
model output changes minimally, contribute disproportionately to the computed
attribution. We propose a variant of IntegratedGradients which primarily
captures gradients in unsaturated regions and evaluate this method on ImageNet
classification networks. We find that this attribution technique shows higher
model faithfulness and lower sensitivity to noise com-pared with standard
Integrated Gradients. A note-book illustrating our computations and results is
available at https://github.com/vivekmig/captum-1/tree/ExpandedIG.
Related papers
- A Learning Paradigm for Interpretable Gradients [9.074325843851726]
We present a novel training approach to improve the quality of gradients for interpretability.
We find that the resulting gradient is qualitatively less noisy and improves quantitatively the interpretability properties of different networks.
arXiv Detail & Related papers (2024-04-23T13:32:29Z) - Neural Gradient Regularizer [150.85797800807524]
We propose a neural gradient regularizer (NGR) that expresses the gradient map as the output of a neural network.
NGR is applicable to various image types and different image processing tasks, functioning in a zero-shot learning fashion.
arXiv Detail & Related papers (2023-08-31T10:19:23Z) - Integrated Decision Gradients: Compute Your Attributions Where the Model
Makes Its Decision [9.385886214196479]
We propose an attribution algorithm called integrated decision gradients (IDG)
IDG focuses on integrating gradients from the region of the path where the model makes its decision, i.e., the portion of the path where the output logit rapidly transitions from zero to its final value.
We minimize the errors within the sum approximation of the path integral by utilizing non-uniform subdivisions determined by adaptive sampling.
arXiv Detail & Related papers (2023-05-31T17:25:12Z) - Deep Manifold Learning with Graph Mining [80.84145791017968]
We propose a novel graph deep model with a non-gradient decision layer for graph mining.
The proposed model has achieved state-of-the-art performance compared to the current models.
arXiv Detail & Related papers (2022-07-18T04:34:08Z) - Geometrically Guided Integrated Gradients [0.3867363075280543]
We introduce an interpretability method called "geometrically-guided integrated gradients"
Our method explores the model's dynamic behavior from multiple scaled versions of the input and captures the best possible attribution for each input.
We also propose a "model perturbation" sanity check to complement the traditionally used "model randomization" test.
arXiv Detail & Related papers (2022-06-13T05:05:43Z) - Gradient Correction beyond Gradient Descent [63.33439072360198]
gradient correction is apparently the most crucial aspect for the training of a neural network.
We introduce a framework (textbfGCGD) to perform gradient correction.
Experiment results show that our gradient correction framework can effectively improve the gradient quality to reduce training epochs by $sim$ 20% and also improve the network performance.
arXiv Detail & Related papers (2022-03-16T01:42:25Z) - On Training Implicit Models [75.20173180996501]
We propose a novel gradient estimate for implicit models, named phantom gradient, that forgoes the costly computation of the exact gradient.
Experiments on large-scale tasks demonstrate that these lightweight phantom gradients significantly accelerate the backward passes in training implicit models by roughly 1.7 times.
arXiv Detail & Related papers (2021-11-09T14:40:24Z) - Integrated Grad-CAM: Sensitivity-Aware Visual Explanation of Deep
Convolutional Networks via Integrated Gradient-Based Scoring [26.434705114982584]
Grad-CAM is a popular solution that provides such a visualization by combining the activation maps obtained from the model.
We introduce a solution to tackle this problem by computing the path integral of the gradient-based terms in Grad-CAM.
We conduct a thorough analysis to demonstrate the improvement achieved by our method in measuring the importance of the extracted representations for the CNN's predictions.
arXiv Detail & Related papers (2021-02-15T19:21:46Z) - A Study of Gradient Variance in Deep Learning [56.437755740715396]
We introduce a method, Gradient Clustering, to minimize the variance of average mini-batch gradient with stratified sampling.
We measure the gradient variance on common deep learning benchmarks and observe that, contrary to common assumptions, gradient variance increases during training.
arXiv Detail & Related papers (2020-07-09T03:23:10Z) - Understanding Integrated Gradients with SmoothTaylor for Deep Neural
Network Attribution [70.78655569298923]
Integrated Gradients as an attribution method for deep neural network models offers simple implementability.
It suffers from noisiness of explanations which affects the ease of interpretability.
The SmoothGrad technique is proposed to solve the noisiness issue and smoothen the attribution maps of any gradient-based attribution method.
arXiv Detail & Related papers (2020-04-22T10:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.