CAManim: Animating end-to-end network activation maps
- URL: http://arxiv.org/abs/2312.11772v1
- Date: Tue, 19 Dec 2023 01:07:36 GMT
- Title: CAManim: Animating end-to-end network activation maps
- Authors: Emily Kaczmarek and Olivier X. Miguel and Alexa C. Bowie and Robin
Ducharme and Alysha L.J. Dingwall-Harvey and Steven Hawken and Christine M.
Armour and Mark C. Walker and Kevin Dick
- Abstract summary: We propose a novel XAI visualization method denoted CAManim that seeks to broaden and focus end-user understanding of CNN predictions.
We additionally propose a novel quantitative assessment that expands upon the Remove and Debias (ROAD) metric.
This builds upon prior research to address the increasing demand for interpretable, robust, and transparent model assessment methodology.
- Score: 0.2509487459755192
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Deep neural networks have been widely adopted in numerous domains due to
their high performance and accessibility to developers and application-specific
end-users. Fundamental to image-based applications is the development of
Convolutional Neural Networks (CNNs), which possess the ability to
automatically extract features from data. However, comprehending these complex
models and their learned representations, which typically comprise millions of
parameters and numerous layers, remains a challenge for both developers and
end-users. This challenge arises due to the absence of interpretable and
transparent tools to make sense of black-box models. There exists a growing
body of Explainable Artificial Intelligence (XAI) literature, including a
collection of methods denoted Class Activation Maps (CAMs), that seek to
demystify what representations the model learns from the data, how it informs a
given prediction, and why it, at times, performs poorly in certain tasks. We
propose a novel XAI visualization method denoted CAManim that seeks to
simultaneously broaden and focus end-user understanding of CNN predictions by
animating the CAM-based network activation maps through all layers, effectively
depicting from end-to-end how a model progressively arrives at the final layer
activation. Herein, we demonstrate that CAManim works with any CAM-based method
and various CNN architectures. Beyond qualitative model assessments, we
additionally propose a novel quantitative assessment that expands upon the
Remove and Debias (ROAD) metric, pairing the qualitative end-to-end network
visual explanations assessment with our novel quantitative "yellow brick ROAD"
assessment (ybROAD). This builds upon prior research to address the increasing
demand for interpretable, robust, and transparent model assessment methodology,
ultimately improving an end-user's trust in a given model's predictions.
Related papers
- CCDepth: A Lightweight Self-supervised Depth Estimation Network with Enhanced Interpretability [11.076431337488973]
This study proposes a novel hybrid self-supervised depth estimation network, CCDepth, comprising convolutional neural networks (CNNs) and the white-box CRATE network.
This novel network uses CNNs and the CRATE modules to extract local and global information in images, respectively, thereby boosting learning efficiency and reducing model size.
arXiv Detail & Related papers (2024-09-30T04:19:40Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Neural Clustering based Visual Representation Learning [61.72646814537163]
Clustering is one of the most classic approaches in machine learning and data analysis.
We propose feature extraction with clustering (FEC), which views feature extraction as a process of selecting representatives from data.
FEC alternates between grouping pixels into individual clusters to abstract representatives and updating the deep features of pixels with current representatives.
arXiv Detail & Related papers (2024-03-26T06:04:50Z) - Data-efficient Large Vision Models through Sequential Autoregression [58.26179273091461]
We develop an efficient, autoregression-based vision model on a limited dataset.
We demonstrate how this model achieves proficiency in a spectrum of visual tasks spanning both high-level and low-level semantic understanding.
Our empirical evaluations underscore the model's agility in adapting to various tasks, heralding a significant reduction in the parameter footprint.
arXiv Detail & Related papers (2024-02-07T13:41:53Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Interpretability of an Interaction Network for identifying $H
\rightarrow b\bar{b}$ jets [4.553120911976256]
In recent times, AI models based on deep neural networks are becoming increasingly popular for many of these applications.
We explore interpretability of AI models by examining an Interaction Network (IN) model designed to identify boosted $Hto bbarb$ jets.
We additionally illustrate the activity of hidden layers within the IN model as Neural Activation Pattern (NAP) diagrams.
arXiv Detail & Related papers (2022-11-23T08:38:52Z) - A Detailed Study of Interpretability of Deep Neural Network based Top
Taggers [3.8541104292281805]
Recent developments in explainable AI (XAI) allow researchers to explore the inner workings of deep neural networks (DNNs)
We explore interpretability of models designed to identify jets coming from top quark decay in high energy proton-proton collisions at the Large Hadron Collider (LHC)
Our studies uncover some major pitfalls of existing XAI methods and illustrate how they can be overcome to obtain consistent and meaningful interpretation of these models.
arXiv Detail & Related papers (2022-10-09T23:02:42Z) - Multi-layer Representation Learning for Robust OOD Image Classification [3.1372269816123994]
We argue that extracting features from a CNN's intermediate layers can assist in the model's final prediction.
Specifically, we adapt the Hypercolumns method to a ResNet-18 and find a significant increase in the model's accuracy, when evaluating on the NICO dataset.
arXiv Detail & Related papers (2022-07-27T17:46:06Z) - Anomaly Detection on Attributed Networks via Contrastive Self-Supervised
Learning [50.24174211654775]
We present a novel contrastive self-supervised learning framework for anomaly detection on attributed networks.
Our framework fully exploits the local information from network data by sampling a novel type of contrastive instance pair.
A graph neural network-based contrastive learning model is proposed to learn informative embedding from high-dimensional attributes and local structure.
arXiv Detail & Related papers (2021-02-27T03:17:20Z) - Learning Deep Interleaved Networks with Asymmetric Co-Attention for
Image Restoration [65.11022516031463]
We present a deep interleaved network (DIN) that learns how information at different states should be combined for high-quality (HQ) images reconstruction.
In this paper, we propose asymmetric co-attention (AsyCA) which is attached at each interleaved node to model the feature dependencies.
Our presented DIN can be trained end-to-end and applied to various image restoration tasks.
arXiv Detail & Related papers (2020-10-29T15:32:00Z) - Eigen-CAM: Class Activation Map using Principal Components [1.2691047660244335]
This paper builds on previous ideas to cope with the increasing demand for interpretable, robust, and transparent models.
The proposed Eigen-CAM computes and visualizes the principle components of the learned features/representations from the convolutional layers.
arXiv Detail & Related papers (2020-08-01T17:14:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.