Cluster-CAM: Cluster-Weighted Visual Interpretation of CNNs' Decision in
Image Classification
- URL: http://arxiv.org/abs/2302.01642v1
- Date: Fri, 3 Feb 2023 10:38:20 GMT
- Title: Cluster-CAM: Cluster-Weighted Visual Interpretation of CNNs' Decision in
Image Classification
- Authors: Zhenpeng Feng, Hongbing Ji, Milos Dakovic, Xiyang Cui, Mingzhe Zhu,
Ljubisa Stankovic
- Abstract summary: Cluster-CAM is an effective and efficient gradient-free CNN interpretation algorithm.
We propose an artful strategy to forge a cognition-base map and cognition-scissors from clustered feature maps.
- Score: 12.971559051829658
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite the tremendous success of convolutional neural networks (CNNs) in
computer vision, the mechanism of CNNs still lacks clear interpretation.
Currently, class activation mapping (CAM), a famous visualization technique to
interpret CNN's decision, has drawn increasing attention. Gradient-based CAMs
are efficient while the performance is heavily affected by gradient vanishing
and exploding. In contrast, gradient-free CAMs can avoid computing gradients to
produce more understandable results. However, existing gradient-free CAMs are
quite time-consuming because hundreds of forward interference per image are
required. In this paper, we proposed Cluster-CAM, an effective and efficient
gradient-free CNN interpretation algorithm. Cluster-CAM can significantly
reduce the times of forward propagation by splitting the feature maps into
clusters in an unsupervised manner. Furthermore, we propose an artful strategy
to forge a cognition-base map and cognition-scissors from clustered feature
maps. The final salience heatmap will be computed by merging the above
cognition maps. Qualitative results conspicuously show that Cluster-CAM can
produce heatmaps where the highlighted regions match the human's cognition more
precisely than existing CAMs. The quantitative evaluation further demonstrates
the superiority of Cluster-CAM in both effectiveness and efficiency.
Related papers
- CAPE: CAM as a Probabilistic Ensemble for Enhanced DNN Interpretation [35.021331140484804]
Class activation maps (CAMs) and recent variants provide ways to visually explain the Deep Neural Networks (DNNs) decision-making process.
We propose CAPE, a novel reformulation of CAM that provides a unified and probabilistically meaningful assessment of the contributions of image regions.
arXiv Detail & Related papers (2024-04-03T01:13:05Z) - BroadCAM: Outcome-agnostic Class Activation Mapping for Small-scale
Weakly Supervised Applications [69.22739434619531]
We propose an outcome-agnostic CAM approach, called BroadCAM, for small-scale weakly supervised applications.
By evaluating BroadCAM on VOC2012 and BCSS-WSSS for WSSS and OpenImages30k for WSOL, BroadCAM demonstrates superior performance.
arXiv Detail & Related papers (2023-09-07T06:45:43Z) - Attention-based Class Activation Diffusion for Weakly-Supervised
Semantic Segmentation [98.306533433627]
extracting class activation maps (CAM) is a key step for weakly-supervised semantic segmentation (WSSS)
This paper proposes a new method to couple CAM and Attention matrix in a probabilistic Diffusion way, and dub it AD-CAM.
Experiments show that AD-CAM as pseudo labels can yield stronger WSSS models than the state-of-the-art variants of CAM.
arXiv Detail & Related papers (2022-11-20T10:06:32Z) - Recipro-CAM: Gradient-free reciprocal class activation map [0.0]
We propose a lightweight architecture and gradient free Reciprocal CAM (Recipro-CAM) to exploit the correlation between activation maps and network outputs.
With the proposed method, we achieved the gains of 1:78 - 3:72% in the ResNet family compared to Score-CAM.
In addition, Recipro-CAM exhibits a saliency map generation rate similar to Grad-CAM and approximately 148 times faster than Score-CAM.
arXiv Detail & Related papers (2022-09-28T13:15:03Z) - VS-CAM: Vertex Semantic Class Activation Mapping to Interpret Vision
Graph Neural Network [10.365366151667017]
Graph convolutional neural network (GCN) has drawn increasing attention and attained good performance in various computer vision tasks.
For standard convolutional neural networks (CNNs), class activation mapping (CAM) methods are commonly used to visualize the connection between CNN's decision and image region by generating a heatmap.
In this paper, we proposed a novel visualization method particularly applicable to GCN, Vertex Semantic Class Activation Mapping (VS-CAM)
arXiv Detail & Related papers (2022-09-15T09:45:59Z) - Shap-CAM: Visual Explanations for Convolutional Neural Networks based on
Shapley Value [86.69600830581912]
We develop a novel visual explanation method called Shap-CAM based on class activation mapping.
We demonstrate that Shap-CAM achieves better visual performance and fairness for interpreting the decision making process.
arXiv Detail & Related papers (2022-08-07T00:59:23Z) - FD-CAM: Improving Faithfulness and Discriminability of Visual
Explanation for CNNs [7.956110316017118]
Class activation map (CAM) has been widely studied for visual explanation of the internal working mechanism of convolutional neural networks.
We propose a novel CAM weighting scheme, named FD-CAM, to improve both the faithfulness and discriminability of the CNN visual explanation.
arXiv Detail & Related papers (2022-06-17T14:08:39Z) - Weakly-supervised fire segmentation by visualizing intermediate CNN
layers [82.75113406937194]
Fire localization in images and videos is an important step for an autonomous system to combat fire incidents.
We consider weakly supervised segmentation of fire in images, in which only image labels are used to train the network.
We show that in the case of fire segmentation, which is a binary segmentation problem, the mean value of features in a mid-layer of classification CNN can perform better than conventional Class Activation Mapping (CAM) method.
arXiv Detail & Related papers (2021-11-16T11:56:28Z) - Towards Learning Spatially Discriminative Feature Representations [26.554140976236052]
We propose a novel loss function, termed as CAM-loss, to constrain the embedded feature maps with the class activation maps (CAMs)
CAM-loss drives the backbone to express the features of target category and suppress the features of non-target categories or background.
Experimental results show that CAM-loss is applicable to a variety of network structures and can be combined with mainstream regularization methods to improve the performance of image classification.
arXiv Detail & Related papers (2021-09-03T08:04:17Z) - TS-CAM: Token Semantic Coupled Attention Map for Weakly Supervised
Object Localization [112.46381729542658]
Weakly supervised object localization (WSOL) is a challenging problem when given image category labels.
We introduce the token semantic coupled attention map (TS-CAM) to take full advantage of the self-attention mechanism in visual transformer for long-range dependency extraction.
arXiv Detail & Related papers (2021-03-27T09:43:16Z) - What Deep CNNs Benefit from Global Covariance Pooling: An Optimization
Perspective [102.37204254403038]
We make an attempt to understand what deep CNNs benefit from GCP in a viewpoint of optimization.
We show that GCP can make the optimization landscape more smooth and the gradients more predictive.
We conduct extensive experiments using various deep CNN models on diversified tasks, and the results provide strong support to our findings.
arXiv Detail & Related papers (2020-03-25T07:00:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.