Understanding CNNs from excitations
- URL: http://arxiv.org/abs/2205.00932v3
- Date: Tue, 16 Jan 2024 02:00:14 GMT
- Title: Understanding CNNs from excitations
- Authors: Zijian Ying, Qianmu Li, Zhichao Lian, Jun Hou, Tong Lin, Tao Wang
- Abstract summary: Saliency maps have proven to be a highly efficacious approach for explicating the decisions of Convolutional Neural Networks.
We present a novel concept, termed positive and negative excitation, which enables the direct extraction of positive and negative excitation for each layer.
- Score: 12.25690353533472
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Saliency maps have proven to be a highly efficacious approach for explicating
the decisions of Convolutional Neural Networks. However, extant methodologies
predominantly rely on gradients, which constrain their ability to explicate
complex models. Furthermore, such approaches are not fully adept at leveraging
negative gradient information to improve interpretive veracity. In this study,
we present a novel concept, termed positive and negative excitation, which
enables the direct extraction of positive and negative excitation for each
layer, thus enabling complete layer-by-layer information utilization sans
gradients. To organize these excitations into final saliency maps, we introduce
a double-chain backpropagation procedure. A comprehensive experimental
evaluation, encompassing both binary classification and multi-classification
tasks, was conducted to gauge the effectiveness of our proposed method.
Encouragingly, the results evince that our approach offers a significant
improvement over the state-of-the-art methods in terms of salient pixel
removal, minor pixel removal, and inconspicuous adversarial perturbation
generation guidance. Additionally, we verify the correlation between positive
and negative excitations.
Related papers
- A Learning Paradigm for Interpretable Gradients [9.074325843851726]
We present a novel training approach to improve the quality of gradients for interpretability.
We find that the resulting gradient is qualitatively less noisy and improves quantitatively the interpretability properties of different networks.
arXiv Detail & Related papers (2024-04-23T13:32:29Z) - Improving Adversarial Transferability via Intermediate-level
Perturbation Decay [79.07074710460012]
We develop a novel intermediate-level method that crafts adversarial examples within a single stage of optimization.
Experimental results show that it outperforms state-of-the-arts by large margins in attacking various victim models.
arXiv Detail & Related papers (2023-04-26T09:49:55Z) - Abs-CAM: A Gradient Optimization Interpretable Approach for Explanation
of Convolutional Neural Networks [7.71412567705588]
Class activation mapping-based method has been widely used to interpret the internal decisions of models in computer vision tasks.
We propose an Absolute value Class Activation Mapping-based (Abs-CAM) method, which optimize the gradients derived from the backpropagation.
The framework of Abs-CAM is divided into two phases: generating initial saliency map and generating final saliency map.
arXiv Detail & Related papers (2022-07-08T02:06:46Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - Incremental False Negative Detection for Contrastive Learning [95.68120675114878]
We introduce a novel incremental false negative detection for self-supervised contrastive learning.
During contrastive learning, we discuss two strategies to explicitly remove the detected false negatives.
Our proposed method outperforms other self-supervised contrastive learning frameworks on multiple benchmarks within a limited compute.
arXiv Detail & Related papers (2021-06-07T15:29:14Z) - Interpreting Deep Neural Networks with Relative Sectional Propagation by
Analyzing Comparative Gradients and Hostile Activations [37.11665902583138]
We propose a new attribution method, Relative Sectional Propagation (RSP), for decomposing the output predictions of Deep Neural Networks (DNNs)
We define hostile factor as an element that interferes with finding the attributions of the target and propagates it in a distinguishable way to overcome the non-suppressed nature of activated neurons.
Our method makes it possible to decompose the predictions of DNNs with clearer class-discriminativeness and detailed elucidations of activation neurons compared to the conventional attribution methods.
arXiv Detail & Related papers (2020-12-07T03:11:07Z) - Rethinking Positive Aggregation and Propagation of Gradients in
Gradient-based Saliency Methods [47.999621481852266]
Saliency methods interpret the prediction of a neural network by showing the importance of input elements for that prediction.
We empirically show that two approaches for handling the gradient information, namely positive aggregation, and positive propagation, break these methods.
arXiv Detail & Related papers (2020-12-01T09:38:54Z) - Understanding Negative Sampling in Graph Representation Learning [87.35038268508414]
We show that negative sampling is as important as positive sampling in determining the optimization objective and the resulted variance.
We propose Metropolis-Hastings (MCNS) to approximate the positive distribution with self-contrast approximation and accelerate negative sampling by Metropolis-Hastings.
We evaluate our method on 5 datasets that cover extensive downstream graph learning tasks, including link prediction, node classification and personalized recommendation.
arXiv Detail & Related papers (2020-05-20T06:25:21Z) - Disentangling Adaptive Gradient Methods from Learning Rates [65.0397050979662]
We take a deeper look at how adaptive gradient methods interact with the learning rate schedule.
We introduce a "grafting" experiment which decouples an update's magnitude from its direction.
We present some empirical and theoretical retrospectives on the generalization of adaptive gradient methods.
arXiv Detail & Related papers (2020-02-26T21:42:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.