Eigen-CAM: Class Activation Map using Principal Components
- URL: http://arxiv.org/abs/2008.00299v1
- Date: Sat, 1 Aug 2020 17:14:13 GMT
- Title: Eigen-CAM: Class Activation Map using Principal Components
- Authors: Mohammed Bany Muhammad, Mohammed Yeasin
- Abstract summary: This paper builds on previous ideas to cope with the increasing demand for interpretable, robust, and transparent models.
The proposed Eigen-CAM computes and visualizes the principle components of the learned features/representations from the convolutional layers.
- Score: 1.2691047660244335
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks are ubiquitous due to the ease of developing models and
their influence on other domains. At the heart of this progress is
convolutional neural networks (CNNs) that are capable of learning
representations or features given a set of data. Making sense of such complex
models (i.e., millions of parameters and hundreds of layers) remains
challenging for developers as well as the end-users. This is partially due to
the lack of tools or interfaces capable of providing interpretability and
transparency. A growing body of literature, for example, class activation map
(CAM), focuses on making sense of what a model learns from the data or why it
behaves poorly in a given task. This paper builds on previous ideas to cope
with the increasing demand for interpretable, robust, and transparent models.
Our approach provides a simpler and intuitive (or familiar) way of generating
CAM. The proposed Eigen-CAM computes and visualizes the principle components of
the learned features/representations from the convolutional layers. Empirical
studies were performed to compare the Eigen-CAM with the state-of-the-art
methods (such as Grad-CAM, Grad-CAM++, CNN-fixations) by evaluating on
benchmark datasets such as weakly-supervised localization and localizing
objects in the presence of adversarial noise. Eigen-CAM was found to be robust
against classification errors made by fully connected layers in CNNs, does not
rely on the backpropagation of gradients, class relevance score, maximum
activation locations, or any other form of weighting features. In addition, it
works with all CNN models without the need to modify layers or retrain models.
Empirical results show up to 12% improvement over the best method among the
methods compared on weakly supervised object localization.
Related papers
- KPCA-CAM: Visual Explainability of Deep Computer Vision Models using Kernel PCA [1.5550533143704957]
This research introduces KPCA-CAM, a technique designed to enhance the interpretability of Convolutional Neural Networks (CNNs)
KPCA-CAM leverages Principal Component Analysis (PCA) with the kernel trick to capture nonlinear relationships within CNN activations more effectively.
Empirical evaluations on the ILSVRC dataset across different CNN models demonstrate that KPCA-CAM produces more precise activation maps.
arXiv Detail & Related papers (2024-09-30T22:36:37Z) - CAManim: Animating end-to-end network activation maps [0.2509487459755192]
We propose a novel XAI visualization method denoted CAManim that seeks to broaden and focus end-user understanding of CNN predictions.
We additionally propose a novel quantitative assessment that expands upon the Remove and Debias (ROAD) metric.
This builds upon prior research to address the increasing demand for interpretable, robust, and transparent model assessment methodology.
arXiv Detail & Related papers (2023-12-19T01:07:36Z) - BroadCAM: Outcome-agnostic Class Activation Mapping for Small-scale
Weakly Supervised Applications [69.22739434619531]
We propose an outcome-agnostic CAM approach, called BroadCAM, for small-scale weakly supervised applications.
By evaluating BroadCAM on VOC2012 and BCSS-WSSS for WSSS and OpenImages30k for WSOL, BroadCAM demonstrates superior performance.
arXiv Detail & Related papers (2023-09-07T06:45:43Z) - Decom--CAM: Tell Me What You See, In Details! Feature-Level Interpretation via Decomposition Class Activation Map [23.71680014689873]
Class Activation Map (CAM) is widely used to interpret deep model predictions by highlighting object location.
This paper proposes a new two-stage interpretability method called the Decomposition Class Activation Map (Decom-CAM)
Our experiments demonstrate that the proposed Decom-CAM outperforms current state-of-the-art methods significantly.
arXiv Detail & Related papers (2023-05-27T14:33:01Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Contrastive Neighborhood Alignment [81.65103777329874]
We present Contrastive Neighborhood Alignment (CNA), a manifold learning approach to maintain the topology of learned features.
The target model aims to mimic the local structure of the source representation space using a contrastive loss.
CNA is illustrated in three scenarios: manifold learning, where the model maintains the local topology of the original data in a dimension-reduced space; model distillation, where a small student model is trained to mimic a larger teacher; and legacy model update, where an older model is replaced by a more powerful one.
arXiv Detail & Related papers (2022-01-06T04:58:31Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.