Fact or Artifact? Revise Layer-wise Relevance Propagation on various ANN
Architectures
- URL: http://arxiv.org/abs/2302.12317v2
- Date: Fri, 30 Jun 2023 17:43:40 GMT
- Title: Fact or Artifact? Revise Layer-wise Relevance Propagation on various ANN
Architectures
- Authors: Marco Landt-Hayen, Willi Rath, Martin Claus and Peer Kr\"oger
- Abstract summary: Layer-wise relevance propagation (LRP) is a powerful technique to reveal insights into various artificial neural network (ANN) architectures.
We show techniques to control model focus and give guidance to improve the quality of obtained relevance maps to separate facts from artifacts.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Layer-wise relevance propagation (LRP) is a widely used and powerful
technique to reveal insights into various artificial neural network (ANN)
architectures. LRP is often used in the context of image classification. The
aim is to understand, which parts of the input sample have highest relevance
and hence most influence on the model prediction. Relevance can be traced back
through the network to attribute a certain score to each input pixel. Relevance
scores are then combined and displayed as heat maps and give humans an
intuitive visual understanding of classification models. Opening the black box
to understand the classification engine in great detail is essential for domain
experts to gain trust in ANN models. However, there are pitfalls in terms of
model-inherent artifacts included in the obtained relevance maps, that can
easily be missed. But for a valid interpretation, these artifacts must not be
ignored. Here, we apply and revise LRP on various ANN architectures trained as
classifiers on geospatial and synthetic data. Depending on the network
architecture, we show techniques to control model focus and give guidance to
improve the quality of obtained relevance maps to separate facts from
artifacts.
Related papers
- Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - Experimental Observations of the Topology of Convolutional Neural
Network Activations [2.4235626091331737]
Topological data analysis provides compact, noise-robust representations of complex structures.
Deep neural networks (DNNs) learn millions of parameters associated with a series of transformations defined by the model architecture.
In this paper, we apply cutting edge techniques from TDA with the goal of gaining insight into the interpretability of convolutional neural networks used for image classification.
arXiv Detail & Related papers (2022-12-01T02:05:44Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Mutual information estimation for graph convolutional neural networks [0.0]
We present an architecture-agnostic method for tracking a network's internal representations during training, which are then used to create a mutual information plane.
We compare how the inductive bias introduced in graph-based architectures changes the mutual information plane relative to a fully connected neural network.
arXiv Detail & Related papers (2022-03-31T08:30:04Z) - Polynomial Networks in Deep Classifiers [55.90321402256631]
We cast the study of deep neural networks under a unifying framework.
Our framework provides insights on the inductive biases of each model.
The efficacy of the proposed models is evaluated on standard image and audio classification benchmarks.
arXiv Detail & Related papers (2021-04-16T06:41:20Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - Eigen-CAM: Class Activation Map using Principal Components [1.2691047660244335]
This paper builds on previous ideas to cope with the increasing demand for interpretable, robust, and transparent models.
The proposed Eigen-CAM computes and visualizes the principle components of the learned features/representations from the convolutional layers.
arXiv Detail & Related papers (2020-08-01T17:14:13Z) - Distilling Knowledge from Graph Convolutional Networks [146.71503336770886]
Existing knowledge distillation methods focus on convolutional neural networks (CNNs)
We propose the first dedicated approach to distilling knowledge from a pre-trained graph convolutional network (GCN) model.
We show that our method achieves the state-of-the-art knowledge distillation performance for GCN models.
arXiv Detail & Related papers (2020-03-23T18:23:11Z) - Explainable Deep Classification Models for Domain Generalization [94.43131722655617]
Explanations are defined as regions of visual evidence upon which a deep classification network makes a decision.
Our training strategy enforces a periodic saliency-based feedback to encourage the model to focus on the image regions that directly correspond to the ground-truth object.
arXiv Detail & Related papers (2020-03-13T22:22:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.