Visual Explanations from Deep Networks via Riemann-Stieltjes Integrated
Gradient-based Localization
- URL: http://arxiv.org/abs/2205.10900v1
- Date: Sun, 22 May 2022 18:30:38 GMT
- Title: Visual Explanations from Deep Networks via Riemann-Stieltjes Integrated
Gradient-based Localization
- Authors: Mirtha Lucas, Miguel Lerma, Jacob Furst, Daniela Raicu
- Abstract summary: We introduce a new technique to produce visual explanations for the predictions of a CNN.
Our method can be applied to any layer of the network, and like Integrated Gradients it is not affected by the problem of vanishing gradients.
Compared to Grad-CAM, heatmaps produced by our algorithm are better focused in the areas of interest, and their numerical computation is more stable.
- Score: 0.24596929878045565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural networks are becoming increasingly better at tasks that involve
classifying and recognizing images. At the same time techniques intended to
explain the network output have been proposed. One such technique is the
Gradient-based Class Activation Map (Grad-CAM), which is able to locate
features of an input image at various levels of a convolutional neural network
(CNN), but is sensitive to the vanishing gradients problem. There are
techniques such as Integrated Gradients (IG), that are not affected by that
problem, but its use is limited to the input layer of a network. Here we
introduce a new technique to produce visual explanations for the predictions of
a CNN. Like Grad-CAM, our method can be applied to any layer of the network,
and like Integrated Gradients it is not affected by the problem of vanishing
gradients. For efficiency, gradient integration is performed numerically at the
layer level using a Riemann-Stieltjes sum approximation. Compared to Grad-CAM,
heatmaps produced by our algorithm are better focused in the areas of interest,
and their numerical computation is more stable. Our code is available at
https://github.com/mlerma54/RSIGradCAM
Related papers
- How to guess a gradient [68.98681202222664]
We show that gradients are more structured than previously thought.
Exploiting this structure can significantly improve gradient-free optimization schemes.
We highlight new challenges in overcoming the large gap between optimizing with exact gradients and guessing the gradients.
arXiv Detail & Related papers (2023-12-07T21:40:44Z) - Rethinking PGD Attack: Is Sign Function Necessary? [131.6894310945647]
We present a theoretical analysis of how such sign-based update algorithm influences step-wise attack performance.
We propose a new raw gradient descent (RGD) algorithm that eliminates the use of sign.
The effectiveness of the proposed RGD algorithm has been demonstrated extensively in experiments.
arXiv Detail & Related papers (2023-12-03T02:26:58Z) - A Framework for Provably Stable and Consistent Training of Deep
Feedforward Networks [4.21061712600981]
We present a novel algorithm for training deep neural networks in supervised (classification and regression) and unsupervised (reinforcement learning) scenarios.
This algorithm combines the standard descent gradient and the gradient clipping method.
We show, in theory and through experiments, that our algorithm updates have low variance, and the training loss reduces in a smooth manner.
arXiv Detail & Related papers (2023-05-20T07:18:06Z) - Gradient Correction beyond Gradient Descent [63.33439072360198]
gradient correction is apparently the most crucial aspect for the training of a neural network.
We introduce a framework (textbfGCGD) to perform gradient correction.
Experiment results show that our gradient correction framework can effectively improve the gradient quality to reduce training epochs by $sim$ 20% and also improve the network performance.
arXiv Detail & Related papers (2022-03-16T01:42:25Z) - Boosting Gradient for White-Box Adversarial Attacks [60.422511092730026]
We propose a universal adversarial example generation method, called ADV-ReLU, to enhance the performance of gradient based white-box attack algorithms.
Our approach calculates the gradient of the loss function versus network input, maps the values to scores, and selects a part of them to update the misleading gradients.
arXiv Detail & Related papers (2020-10-21T02:13:26Z) - Scaling Symbolic Methods using Gradients for Neural Model Explanation [22.568591780291776]
We propose a technique for combining gradient-based methods with symbolic techniques to scale such analyses.
In particular, we apply this technique to identify minimal regions in an input that are most relevant for a neural network's prediction.
We evaluate our technique on three datasets - MNIST, ImageNet, and Beer Reviews.
arXiv Detail & Related papers (2020-06-29T19:12:22Z) - GradAug: A New Regularization Method for Deep Neural Networks [19.239311087570318]
We propose a new regularization method to alleviate over-fitting in deep neural networks.
The proposed method introduces self-guided disturbances to the raw gradients of the network.
We demonstrate that GradAug can help the network learn well-generalized and more diverse representations.
arXiv Detail & Related papers (2020-06-14T19:30:34Z) - Understanding Integrated Gradients with SmoothTaylor for Deep Neural
Network Attribution [70.78655569298923]
Integrated Gradients as an attribution method for deep neural network models offers simple implementability.
It suffers from noisiness of explanations which affects the ease of interpretability.
The SmoothGrad technique is proposed to solve the noisiness issue and smoothen the attribution maps of any gradient-based attribution method.
arXiv Detail & Related papers (2020-04-22T10:43:19Z) - Saliency Enhancement using Gradient Domain Edges Merging [65.90255950853674]
We develop a method to merge the edges with the saliency maps to improve the performance of the saliency.
This leads to our proposed saliency enhancement using edges (SEE) with an average improvement of at least 3.4 times higher on the DUT-OMRON dataset.
The SEE algorithm is split into 2 parts, SEE-Pre for preprocessing and SEE-Post pour postprocessing.
arXiv Detail & Related papers (2020-02-11T14:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.