Empowering CAM-Based Methods with Capability to Generate Fine-Grained
and High-Faithfulness Explanations
- URL: http://arxiv.org/abs/2303.09171v3
- Date: Wed, 31 Jan 2024 06:27:21 GMT
- Title: Empowering CAM-Based Methods with Capability to Generate Fine-Grained
and High-Faithfulness Explanations
- Authors: Changqing Qiu, Fusheng Jin, Yining Zhang
- Abstract summary: We propose FG-CAM, which extends CAM-based methods to enable generating fine-grained and high-faithfulness explanations.
Our method not only solves the shortcoming of CAM-based methods without changing their characteristics, but also generates fine-grained explanations that have higher faithfulness than LRP and its variants.
- Score: 1.757194730633422
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, the explanation of neural network models has garnered considerable
research attention. In computer vision, CAM (Class Activation Map)-based
methods and LRP (Layer-wise Relevance Propagation) method are two common
explanation methods. However, since most CAM-based methods can only generate
global weights, they can only generate coarse-grained explanations at a deep
layer. LRP and its variants, on the other hand, can generate fine-grained
explanations. But the faithfulness of the explanations is too low. To address
these challenges, in this paper, we propose FG-CAM (Fine-Grained CAM), which
extends CAM-based methods to enable generating fine-grained and
high-faithfulness explanations. FG-CAM uses the relationship between two
adjacent layers of feature maps with resolution differences to gradually
increase the explanation resolution, while finding the contributing pixels and
filtering out the pixels that do not contribute. Our method not only solves the
shortcoming of CAM-based methods without changing their characteristics, but
also generates fine-grained explanations that have higher faithfulness than LRP
and its variants. We also present FG-CAM with denoising, which is a variant of
FG-CAM and is able to generate less noisy explanations with almost no change in
explanation faithfulness. Experimental results show that the performance of
FG-CAM is almost unaffected by the explanation resolution. FG-CAM outperforms
existing CAM-based methods significantly in both shallow and intermediate
layers, and outperforms LRP and its variants significantly in the input layer.
Our code is available at https://github.com/dongmo-qcq/FG-CAM.
Related papers
- Enhancing Explainable AI: A Hybrid Approach Combining GradCAM and LRP for CNN Interpretability [0.0]
We present a new technique that explains the output of a CNN-based model using a combination of GradCAM and LRP methods.
Both of these methods produce visual explanations by highlighting input regions that are important for predictions.
arXiv Detail & Related papers (2024-05-20T16:58:24Z) - BroadCAM: Outcome-agnostic Class Activation Mapping for Small-scale
Weakly Supervised Applications [69.22739434619531]
We propose an outcome-agnostic CAM approach, called BroadCAM, for small-scale weakly supervised applications.
By evaluating BroadCAM on VOC2012 and BCSS-WSSS for WSSS and OpenImages30k for WSOL, BroadCAM demonstrates superior performance.
arXiv Detail & Related papers (2023-09-07T06:45:43Z) - Exploit CAM by itself: Complementary Learning System for Weakly
Supervised Semantic Segmentation [59.24824050194334]
This paper turns to an interesting working mechanism in agent learning named Complementary Learning System ( CLS)
Motivated by this simple but effective learning pattern, we propose a General-Specific Learning Mechanism (GSLM)
GSLM develops a General Learning Module (GLM) and a Specific Learning Module (SLM)
arXiv Detail & Related papers (2023-03-04T16:16:47Z) - Attention-based Class Activation Diffusion for Weakly-Supervised
Semantic Segmentation [98.306533433627]
extracting class activation maps (CAM) is a key step for weakly-supervised semantic segmentation (WSSS)
This paper proposes a new method to couple CAM and Attention matrix in a probabilistic Diffusion way, and dub it AD-CAM.
Experiments show that AD-CAM as pseudo labels can yield stronger WSSS models than the state-of-the-art variants of CAM.
arXiv Detail & Related papers (2022-11-20T10:06:32Z) - Recipro-CAM: Gradient-free reciprocal class activation map [0.0]
We propose a lightweight architecture and gradient free Reciprocal CAM (Recipro-CAM) to exploit the correlation between activation maps and network outputs.
With the proposed method, we achieved the gains of 1:78 - 3:72% in the ResNet family compared to Score-CAM.
In addition, Recipro-CAM exhibits a saliency map generation rate similar to Grad-CAM and approximately 148 times faster than Score-CAM.
arXiv Detail & Related papers (2022-09-28T13:15:03Z) - FD-CAM: Improving Faithfulness and Discriminability of Visual
Explanation for CNNs [7.956110316017118]
Class activation map (CAM) has been widely studied for visual explanation of the internal working mechanism of convolutional neural networks.
We propose a novel CAM weighting scheme, named FD-CAM, to improve both the faithfulness and discriminability of the CNN visual explanation.
arXiv Detail & Related papers (2022-06-17T14:08:39Z) - Class Re-Activation Maps for Weakly-Supervised Semantic Segmentation [88.55040177178442]
Class activation maps (CAM) is arguably the most standard step of generating pseudo masks for semantic segmentation.
Yet, the crux of the unsatisfactory pseudo masks is the binary cross-entropy loss (BCE) widely used in CAM.
We introduce an embarrassingly simple yet surprisingly effective method: Reactivating the converged CAM with BCE by using softmax cross-entropy loss (SCE)
The evaluation on both PASCAL VOC and MSCOCO shows that ReCAM not only generates high-quality masks, but also supports plug-and-play in any CAM variant with little overhead.
arXiv Detail & Related papers (2022-03-02T09:14:58Z) - PCAM: Product of Cross-Attention Matrices for Rigid Registration of
Point Clouds [79.99653758293277]
PCAM is a neural network whose key element is a pointwise product of cross-attention matrices.
We show that PCAM achieves state-of-the-art results among methods which, like us, solve steps (a) and (b) jointly via deepnets.
arXiv Detail & Related papers (2021-10-04T09:23:27Z) - F-CAM: Full Resolution CAM via Guided Parametric Upscaling [20.609010268320013]
Class Activation Mapping (CAM) methods have recently gained much attention for weakly-supervised object localization (WSOL) tasks.
CAM methods are typically integrated within off-the-shelf CNN backbones, such as ResNet50.
We introduce a generic method for parametric upscaling of CAMs that allows constructing accurate full resolution CAMs.
arXiv Detail & Related papers (2021-09-15T04:45:20Z) - Use HiResCAM instead of Grad-CAM for faithful explanations of
convolutional neural networks [89.56292219019163]
Explanation methods facilitate the development of models that learn meaningful concepts and avoid exploiting spurious correlations.
We illustrate a previously unrecognized limitation of the popular neural network explanation method Grad-CAM.
We propose HiResCAM, a class-specific explanation method that is guaranteed to highlight only the locations the model used to make each prediction.
arXiv Detail & Related papers (2020-11-17T19:26:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.