Rethinking gradient weights' influence over saliency map estimation
- URL: http://arxiv.org/abs/2207.05374v1
- Date: Tue, 12 Jul 2022 08:14:57 GMT
- Title: Rethinking gradient weights' influence over saliency map estimation
- Authors: Masud An Nur Islam Fahim, Nazmus Saqib, Shafkat Khan Siam, Ho Yub Jung
- Abstract summary: Class activation map (CAM) helps to formulate saliency maps that aid in interpreting the deep neural network's prediction.
We propose a global guidance map to rectify the weighted aggregation operation during saliency estimation.
The proposed scheme achieves significant improvement over the test images from the ImageNet, MS-COCO 14, and PASCAL VOC 2012 datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Class activation map (CAM) helps to formulate saliency maps that aid in
interpreting the deep neural network's prediction. Gradient-based methods are
generally faster than other branches of vision interpretability and independent
of human guidance. The performance of CAM-like studies depends on the governing
model's layer response, and the influences of the gradients. Typical
gradient-oriented CAM studies rely on weighted aggregation for saliency map
estimation by projecting the gradient maps into single weight values, which may
lead to over generalized saliency map. To address this issue, we use a global
guidance map to rectify the weighted aggregation operation during saliency
estimation, where resultant interpretations are comparatively clean er and
instance-specific. We obtain the global guidance map by performing elementwise
multiplication between the feature maps and their corresponding gradient maps.
To validate our study, we compare the proposed study with eight different
saliency visualizers. In addition, we use seven commonly used evaluation
metrics for quantitative comparison. The proposed scheme achieves significant
improvement over the test images from the ImageNet, MS-COCO 14, and PASCAL VOC
2012 datasets.
Related papers
- Dynamic Decoupling of Placid Terminal Attractor-based Gradient Descent Algorithm [56.06235614890066]
Gradient descent (GD) and gradient descent (SGD) have been widely used in a number of application domains.
This paper carefully analyzes the dynamics of GD based on the terminal attractor at different stages of its gradient flow.
arXiv Detail & Related papers (2024-09-10T14:15:56Z) - Guided AbsoluteGrad: Magnitude of Gradients Matters to Explanation's Localization and Saliency [10.786952260623002]
We propose a gradient-based XAI method called Guided AbsoluteGrad for saliency map explanations.
We introduce a novel evaluation metric named ReCover And Predict (RCAP), which considers the localization and visual noise level objectives.
We evaluate Guided AbsoluteGrad with seven gradient-based XAI methods using the RCAP metric and other SOTA metrics in three case studies.
arXiv Detail & Related papers (2024-04-23T23:26:02Z) - A Learning Paradigm for Interpretable Gradients [9.074325843851726]
We present a novel training approach to improve the quality of gradients for interpretability.
We find that the resulting gradient is qualitatively less noisy and improves quantitatively the interpretability properties of different networks.
arXiv Detail & Related papers (2024-04-23T13:32:29Z) - Neural Gradient Regularizer [150.85797800807524]
We propose a neural gradient regularizer (NGR) that expresses the gradient map as the output of a neural network.
NGR is applicable to various image types and different image processing tasks, functioning in a zero-shot learning fashion.
arXiv Detail & Related papers (2023-08-31T10:19:23Z) - Interpretation of Neural Networks is Susceptible to Universal Adversarial Perturbations [9.054540533394926]
We show the existence of a Universal Perturbation for Interpretation (UPI) for standard image datasets.
We propose a gradient-based optimization method as well as a principal component analysis (PCA)-based approach to compute a UPI which can effectively alter a neural network's gradient-based interpretation on different samples.
arXiv Detail & Related papers (2022-11-30T15:55:40Z) - Abs-CAM: A Gradient Optimization Interpretable Approach for Explanation
of Convolutional Neural Networks [7.71412567705588]
Class activation mapping-based method has been widely used to interpret the internal decisions of models in computer vision tasks.
We propose an Absolute value Class Activation Mapping-based (Abs-CAM) method, which optimize the gradients derived from the backpropagation.
The framework of Abs-CAM is divided into two phases: generating initial saliency map and generating final saliency map.
arXiv Detail & Related papers (2022-07-08T02:06:46Z) - TSG: Target-Selective Gradient Backprop for Probing CNN Visual Saliency [72.9106103283475]
We study the visual saliency, a.k.a. visual explanation, to interpret convolutional neural networks.
Inspired by those observations, we propose a novel visual saliency framework, termed Target-Selective Gradient (TSG) backprop.
The proposed TSG consists of two components, namely, TSG-Conv and TSG-FC, which rectify the gradients for convolutional layers and fully-connected layers, respectively.
arXiv Detail & Related papers (2021-10-11T12:00:20Z) - CAMERAS: Enhanced Resolution And Sanity preserving Class Activation
Mapping for image saliency [61.40511574314069]
Backpropagation image saliency aims at explaining model predictions by estimating model-centric importance of individual pixels in the input.
We propose CAMERAS, a technique to compute high-fidelity backpropagation saliency maps without requiring any external priors.
arXiv Detail & Related papers (2021-06-20T08:20:56Z) - Rethinking Localization Map: Towards Accurate Object Perception with
Self-Enhancement Maps [78.2581910688094]
This work introduces a novel self-enhancement method to harvest accurate object localization maps and object boundaries with only category labels as supervision.
In particular, the proposed Self-Enhancement Maps achieve the state-of-the-art localization accuracy of 54.88% on ILSVRC.
arXiv Detail & Related papers (2020-06-09T12:35:55Z) - Understanding Integrated Gradients with SmoothTaylor for Deep Neural
Network Attribution [70.78655569298923]
Integrated Gradients as an attribution method for deep neural network models offers simple implementability.
It suffers from noisiness of explanations which affects the ease of interpretability.
The SmoothGrad technique is proposed to solve the noisiness issue and smoothen the attribution maps of any gradient-based attribution method.
arXiv Detail & Related papers (2020-04-22T10:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.