Axiom-based Grad-CAM: Towards Accurate Visualization and Explanation of
CNNs
- URL: http://arxiv.org/abs/2008.02312v4
- Date: Wed, 19 Aug 2020 06:04:28 GMT
- Title: Axiom-based Grad-CAM: Towards Accurate Visualization and Explanation of
CNNs
- Authors: Ruigang Fu, Qingyong Hu, Xiaohu Dong, Yulan Guo, Yinghui Gao, Biao Li
- Abstract summary: Class Activation Mapping (CAM) methods have been proposed to discover the connection between CNN's decision and image regions.
In this paper, we introduce two axioms -- Conservation and Sensitivity -- to the visualization paradigm of the CAM methods.
A dedicated Axiom-based Grad-CAM (XGrad-CAM) is proposed to satisfy these axioms as much as possible.
- Score: 29.731732363623713
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: To have a better understanding and usage of Convolution Neural Networks
(CNNs), the visualization and interpretation of CNNs has attracted increasing
attention in recent years. In particular, several Class Activation Mapping
(CAM) methods have been proposed to discover the connection between CNN's
decision and image regions. In spite of the reasonable visualization, lack of
clear and sufficient theoretical support is the main limitation of these
methods. In this paper, we introduce two axioms -- Conservation and Sensitivity
-- to the visualization paradigm of the CAM methods. Meanwhile, a dedicated
Axiom-based Grad-CAM (XGrad-CAM) is proposed to satisfy these axioms as much as
possible. Experiments demonstrate that XGrad-CAM is an enhanced version of
Grad-CAM in terms of conservation and sensitivity. It is able to achieve better
visualization performance than Grad-CAM, while also be class-discriminative and
easy-to-implement compared with Grad-CAM++ and Ablation-CAM. The code is
available at https://github.com/Fu0511/XGrad-CAM.
Related papers
- Generalizing GradCAM for Embedding Networks [0.0]
We present a new method EmbeddingCAM, which generalizes the Grad-CAM for embedding networks.
We show the effectiveness of our method on CUB-200-2011 dataset and also present quantitative and qualitative analysis on the dataset.
arXiv Detail & Related papers (2024-02-01T04:58:06Z) - FM-G-CAM: A Holistic Approach for Explainable AI in Computer Vision [0.6215404942415159]
We emphasise the need to understand predictions of Computer Vision models, specifically Convolutional Neural Network (CNN) based models.
Existing methods of explaining CNN predictions are mostly based on Gradient-weighted Class Activation Maps (Grad-CAM) and solely focus on a single target class.
We present an exhaustive methodology called Fused Multi-class Gradient-weighted Class Activation Map (FM-G-CAM) that considers multiple top predicted classes.
arXiv Detail & Related papers (2023-12-10T19:33:40Z) - BroadCAM: Outcome-agnostic Class Activation Mapping for Small-scale
Weakly Supervised Applications [69.22739434619531]
We propose an outcome-agnostic CAM approach, called BroadCAM, for small-scale weakly supervised applications.
By evaluating BroadCAM on VOC2012 and BCSS-WSSS for WSSS and OpenImages30k for WSOL, BroadCAM demonstrates superior performance.
arXiv Detail & Related papers (2023-09-07T06:45:43Z) - Exploit CAM by itself: Complementary Learning System for Weakly
Supervised Semantic Segmentation [59.24824050194334]
This paper turns to an interesting working mechanism in agent learning named Complementary Learning System ( CLS)
Motivated by this simple but effective learning pattern, we propose a General-Specific Learning Mechanism (GSLM)
GSLM develops a General Learning Module (GLM) and a Specific Learning Module (SLM)
arXiv Detail & Related papers (2023-03-04T16:16:47Z) - Cluster-CAM: Cluster-Weighted Visual Interpretation of CNNs' Decision in
Image Classification [12.971559051829658]
Cluster-CAM is an effective and efficient gradient-free CNN interpretation algorithm.
We propose an artful strategy to forge a cognition-base map and cognition-scissors from clustered feature maps.
arXiv Detail & Related papers (2023-02-03T10:38:20Z) - Attention-based Class Activation Diffusion for Weakly-Supervised
Semantic Segmentation [98.306533433627]
extracting class activation maps (CAM) is a key step for weakly-supervised semantic segmentation (WSSS)
This paper proposes a new method to couple CAM and Attention matrix in a probabilistic Diffusion way, and dub it AD-CAM.
Experiments show that AD-CAM as pseudo labels can yield stronger WSSS models than the state-of-the-art variants of CAM.
arXiv Detail & Related papers (2022-11-20T10:06:32Z) - FD-CAM: Improving Faithfulness and Discriminability of Visual
Explanation for CNNs [7.956110316017118]
Class activation map (CAM) has been widely studied for visual explanation of the internal working mechanism of convolutional neural networks.
We propose a novel CAM weighting scheme, named FD-CAM, to improve both the faithfulness and discriminability of the CNN visual explanation.
arXiv Detail & Related papers (2022-06-17T14:08:39Z) - Class Re-Activation Maps for Weakly-Supervised Semantic Segmentation [88.55040177178442]
Class activation maps (CAM) is arguably the most standard step of generating pseudo masks for semantic segmentation.
Yet, the crux of the unsatisfactory pseudo masks is the binary cross-entropy loss (BCE) widely used in CAM.
We introduce an embarrassingly simple yet surprisingly effective method: Reactivating the converged CAM with BCE by using softmax cross-entropy loss (SCE)
The evaluation on both PASCAL VOC and MSCOCO shows that ReCAM not only generates high-quality masks, but also supports plug-and-play in any CAM variant with little overhead.
arXiv Detail & Related papers (2022-03-02T09:14:58Z) - Towards Learning Spatially Discriminative Feature Representations [26.554140976236052]
We propose a novel loss function, termed as CAM-loss, to constrain the embedded feature maps with the class activation maps (CAMs)
CAM-loss drives the backbone to express the features of target category and suppress the features of non-target categories or background.
Experimental results show that CAM-loss is applicable to a variety of network structures and can be combined with mainstream regularization methods to improve the performance of image classification.
arXiv Detail & Related papers (2021-09-03T08:04:17Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Use HiResCAM instead of Grad-CAM for faithful explanations of
convolutional neural networks [89.56292219019163]
Explanation methods facilitate the development of models that learn meaningful concepts and avoid exploiting spurious correlations.
We illustrate a previously unrecognized limitation of the popular neural network explanation method Grad-CAM.
We propose HiResCAM, a class-specific explanation method that is guaranteed to highlight only the locations the model used to make each prediction.
arXiv Detail & Related papers (2020-11-17T19:26:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.