Optimizing Explanations by Network Canonization and Hyperparameter
Search
- URL: http://arxiv.org/abs/2211.17174v2
- Date: Mon, 27 Mar 2023 09:42:13 GMT
- Title: Optimizing Explanations by Network Canonization and Hyperparameter
Search
- Authors: Frederik Pahde, Galip \"Umit Yolcu, Alexander Binder, Wojciech Samek,
Sebastian Lapuschkin
- Abstract summary: Rule-based and modified backpropagation XAI approaches often face challenges when being applied to modern model architectures.
Model canonization is the process of re-structuring the model to disregard problematic components without changing the underlying function.
In this work, we propose canonizations for currently relevant model blocks applicable to popular deep neural network architectures.
- Score: 74.76732413972005
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Explainable AI (XAI) is slowly becoming a key component for many AI
applications. Rule-based and modified backpropagation XAI approaches however
often face challenges when being applied to modern model architectures
including innovative layer building blocks, which is caused by two reasons.
Firstly, the high flexibility of rule-based XAI methods leads to numerous
potential parameterizations. Secondly, many XAI methods break the
implementation-invariance axiom because they struggle with certain model
components, e.g., BatchNorm layers. The latter can be addressed with model
canonization, which is the process of re-structuring the model to disregard
problematic components without changing the underlying function. While model
canonization is straightforward for simple architectures (e.g., VGG, ResNet),
it can be challenging for more complex and highly interconnected models (e.g.,
DenseNet). Moreover, there is only little quantifiable evidence that model
canonization is beneficial for XAI. In this work, we propose canonizations for
currently relevant model blocks applicable to popular deep neural network
architectures,including VGG, ResNet, EfficientNet, DenseNets, as well as
Relation Networks. We further suggest a XAI evaluation framework with which we
quantify and compare the effect sof model canonization for various XAI methods
in image classification tasks on the Pascal-VOC and ILSVRC2017 datasets, as
well as for Visual Question Answering using CLEVR-XAI. Moreover, addressing the
former issue outlined above, we demonstrate how our evaluation framework can be
applied to perform hyperparameter search for XAI methods to optimize the
quality of explanations.
Related papers
- Explainable AI for Enhancing Efficiency of DL-based Channel Estimation [1.0136215038345013]
Support of artificial intelligence based decision-making is a key element in future 6G networks.
In such applications, using AI as black-box models is risky and challenging.
We propose a novel-based XAI-CHEST framework that is oriented toward channel estimation in wireless communications.
arXiv Detail & Related papers (2024-07-09T16:24:21Z) - Extending CAM-based XAI methods for Remote Sensing Imagery Segmentation [7.735470452949379]
We introduce a new XAI evaluation methodology and metric based on "Entropy" to measure the model uncertainty.
We show that using Entropy to monitor the model uncertainty in segmenting the pixels within the target class is more suitable.
arXiv Detail & Related papers (2023-10-03T07:01:23Z) - Trainable Noise Model as an XAI evaluation method: application on Sobol
for remote sensing image segmentation [0.5735035463793009]
This paper adapts the gradient-free Sobol XAI method for semantic segmentation.
A benchmark analysis is conducted to evaluate and compare performance of three XAI methods.
arXiv Detail & Related papers (2023-10-03T06:51:48Z) - REX: Rapid Exploration and eXploitation for AI Agents [103.68453326880456]
We propose an enhanced approach for Rapid Exploration and eXploitation for AI Agents called REX.
REX introduces an additional layer of rewards and integrates concepts similar to Upper Confidence Bound (UCB) scores, leading to more robust and efficient AI agent performance.
arXiv Detail & Related papers (2023-07-18T04:26:33Z) - Re-parameterizing Your Optimizers rather than Architectures [119.08740698936633]
We propose a novel paradigm of incorporating model-specific prior knowledge into Structurals and using them to train generic (simple) models.
As an implementation, we propose a novel methodology to add prior knowledge by modifying the gradients according to a set of model-specific hyper- parameters.
For a simple model trained with a Repr, we focus on a VGG-style plain model and showcase that such a simple model trained with a Repr, which is referred to as Rep-VGG, performs on par with the recent well-designed models.
arXiv Detail & Related papers (2022-05-30T16:55:59Z) - Dynamically-Scaled Deep Canonical Correlation Analysis [77.34726150561087]
Canonical Correlation Analysis (CCA) is a method for feature extraction of two views by finding maximally correlated linear projections of them.
We introduce a novel dynamic scaling method for training an input-dependent canonical correlation model.
arXiv Detail & Related papers (2022-03-23T12:52:49Z) - Interpretable pipelines with evolutionarily optimized modules for RL
tasks with visual inputs [5.254093731341154]
We propose end-to-end pipelines composed of multiple interpretable models co-optimized by means of evolutionary algorithms.
We test our approach in reinforcement learning environments from the Atari benchmark.
arXiv Detail & Related papers (2022-02-10T10:33:44Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Fast and Robust Cascade Model for Multiple Degradation Single Image
Super-Resolution [2.1574781022415364]
Single Image Super-Resolution (SISR) is one of the low-level computer vision problems that has received increased attention in the last few years.
Here, we propose a new formulation of the Convolutional Neural Network (CNN) cascade model.
A new densely connected CNN-architecture is proposed where the output of each sub- module is restricted using some external knowledge.
arXiv Detail & Related papers (2020-11-16T18:59:49Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.