Input Bias in Rectified Gradients and Modified Saliency Maps
- URL: http://arxiv.org/abs/2011.05002v3
- Date: Tue, 1 Dec 2020 10:34:25 GMT
- Title: Input Bias in Rectified Gradients and Modified Saliency Maps
- Authors: Lennart Brocki, Neo Christopher Chung
- Abstract summary: Saliency maps provide an intuitive way to identify input features with substantial influences on classifications or latent concepts.
Several modifications to conventional saliency maps, such as Rectified Gradients, have been introduced to allegedly denoise and improve interpretability.
We demonstrate that dark areas of an input image are not highlighted by a saliency map using Rectified Gradients, even if it is relevant for the class or concept.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Interpretation and improvement of deep neural networks relies on better
understanding of their underlying mechanisms. In particular, gradients of
classes or concepts with respect to the input features (e.g., pixels in images)
are often used as importance scores or estimators, which are visualized in
saliency maps. Thus, a family of saliency methods provide an intuitive way to
identify input features with substantial influences on classifications or
latent concepts. Several modifications to conventional saliency maps, such as
Rectified Gradients and Layer-wise Relevance Propagation (LRP), have been
introduced to allegedly denoise and improve interpretability. While visually
coherent in certain cases, Rectified Gradients and other modified saliency maps
introduce a strong input bias (e.g., brightness in the RGB space) because of
inappropriate uses of the input features. We demonstrate that dark areas of an
input image are not highlighted by a saliency map using Rectified Gradients,
even if it is relevant for the class or concept. Even in the scaled images, the
input bias exists around an artificial point in color spectrum. Our
modification, which simply eliminates multiplication with input features,
removes this bias. This showcases how a visual criteria may not align with true
explainability of deep learning models.
Related papers
- Unlearning-based Neural Interpretations [51.99182464831169]
We show that current baselines defined using static functions are biased, fragile and manipulable.
We propose UNI to compute an (un)learnable, debiased and adaptive baseline by perturbing the input towards an unlearning direction of steepest ascent.
arXiv Detail & Related papers (2024-10-10T16:02:39Z) - A Learning Paradigm for Interpretable Gradients [9.074325843851726]
We present a novel training approach to improve the quality of gradients for interpretability.
We find that the resulting gradient is qualitatively less noisy and improves quantitatively the interpretability properties of different networks.
arXiv Detail & Related papers (2024-04-23T13:32:29Z) - Classes Are Not Equal: An Empirical Study on Image Recognition Fairness [100.36114135663836]
We experimentally demonstrate that classes are not equal and the fairness issue is prevalent for image classification models across various datasets.
Our findings reveal that models tend to exhibit greater prediction biases for classes that are more challenging to recognize.
Data augmentation and representation learning algorithms improve overall performance by promoting fairness to some degree in image classification.
arXiv Detail & Related papers (2024-02-28T07:54:50Z) - Improving Fairness using Vision-Language Driven Image Augmentation [60.428157003498995]
Fairness is crucial when training a deep-learning discriminative model, especially in the facial domain.
Models tend to correlate specific characteristics (such as age and skin color) with unrelated attributes (downstream tasks)
This paper proposes a method to mitigate these correlations to improve fairness.
arXiv Detail & Related papers (2023-11-02T19:51:10Z) - Neural Gradient Regularizer [150.85797800807524]
We propose a neural gradient regularizer (NGR) that expresses the gradient map as the output of a neural network.
NGR is applicable to various image types and different image processing tasks, functioning in a zero-shot learning fashion.
arXiv Detail & Related papers (2023-08-31T10:19:23Z) - CAMERAS: Enhanced Resolution And Sanity preserving Class Activation
Mapping for image saliency [61.40511574314069]
Backpropagation image saliency aims at explaining model predictions by estimating model-centric importance of individual pixels in the input.
We propose CAMERAS, a technique to compute high-fidelity backpropagation saliency maps without requiring any external priors.
arXiv Detail & Related papers (2021-06-20T08:20:56Z) - Evaluating Input Perturbation Methods for Interpreting CNNs and Saliency
Map Comparison [9.023847175654602]
In this paper we show that arguably neutral baseline images still impact the generated saliency maps and their evaluation with input perturbations.
We experimentally reveal inconsistencies among a selection of input perturbation methods and find that they lack robustness for generating saliency maps and for evaluating saliency maps as saliency metrics.
arXiv Detail & Related papers (2021-01-26T18:11:06Z) - Understanding Integrated Gradients with SmoothTaylor for Deep Neural
Network Attribution [70.78655569298923]
Integrated Gradients as an attribution method for deep neural network models offers simple implementability.
It suffers from noisiness of explanations which affects the ease of interpretability.
The SmoothGrad technique is proposed to solve the noisiness issue and smoothen the attribution maps of any gradient-based attribution method.
arXiv Detail & Related papers (2020-04-22T10:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.