Explaining Neural Networks by Decoding Layer Activations
- URL: http://arxiv.org/abs/2005.13630v3
- Date: Fri, 26 Feb 2021 17:11:43 GMT
- Title: Explaining Neural Networks by Decoding Layer Activations
- Authors: Johannes Schneider and Michalis Vlachos
- Abstract summary: We present a CLAssifier-DECoder' architecture (emphClaDec) which facilitates the comprehension of the output of an arbitrary layer in a neural network (NN)
It uses a decoder to transform the non-interpretable representation of the given layer to a representation more similar to the domain a human is familiar with.
In an image recognition problem, one can recognize what information is represented by a layer by contrasting reconstructed images of emphClaDec with those of a conventional auto-encoder(AE) serving as reference.
- Score: 3.6245632117657816
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a `CLAssifier-DECoder' architecture (\emph{ClaDec}) which
facilitates the comprehension of the output of an arbitrary layer in a neural
network (NN). It uses a decoder to transform the non-interpretable
representation of the given layer to a representation that is more similar to
the domain a human is familiar with. In an image recognition problem, one can
recognize what information is represented by a layer by contrasting
reconstructed images of \emph{ClaDec} with those of a conventional
auto-encoder(AE) serving as reference. We also extend \emph{ClaDec} to allow
the trade-off between human interpretability and fidelity. We evaluate our
approach for image classification using Convolutional NNs. We show that
reconstructed visualizations using encodings from a classifier capture more
relevant information for classification than conventional AEs. Relevant code is
available at \url{https://github.com/JohnTailor/ClaDec}
Related papers
- Spatial Action Unit Cues for Interpretable Deep Facial Expression Recognition [55.97779732051921]
State-of-the-art classifiers for facial expression recognition (FER) lack interpretability, an important feature for end-users.
A new learning strategy is proposed to explicitly incorporate AU cues into classifier training, allowing to train deep interpretable models.
Our new strategy is generic, and can be applied to any deep CNN- or transformer-based classifier without requiring any architectural change or significant additional training time.
arXiv Detail & Related papers (2024-10-01T10:42:55Z) - Guided Interpretable Facial Expression Recognition via Spatial Action Unit Cues [55.97779732051921]
A new learning strategy is proposed to explicitly incorporate au cues into classifier training.
We show that our strategy can improve layer-wise interpretability without degrading classification performance.
arXiv Detail & Related papers (2024-02-01T02:13:49Z) - An Image captioning algorithm based on the Hybrid Deep Learning
Technique (CNN+GRU) [0.0]
We present a CNN-GRU encoder decode framework for caption-to-image reconstructor.
It handles the semantic context into consideration as well as the time complexity.
The suggested model outperforms the state-of-the-art LSTM-A5 model for picture captioning in terms of time complexity and accuracy.
arXiv Detail & Related papers (2023-01-06T10:00:06Z) - Traditional Classification Neural Networks are Good Generators: They are
Competitive with DDPMs and GANs [104.72108627191041]
We show that conventional neural network classifiers can generate high-quality images comparable to state-of-the-art generative models.
We propose a mask-based reconstruction module to make semantic gradients-aware to synthesize plausible images.
We show that our method is also applicable to text-to-image generation by regarding image-text foundation models.
arXiv Detail & Related papers (2022-11-27T11:25:35Z) - Towards Semantic Communications: Deep Learning-Based Image Semantic
Coding [42.453963827153856]
We conceive the semantic communications for image data that is much more richer in semantics and bandwidth sensitive.
We propose an reinforcement learning based adaptive semantic coding (RL-ASC) approach that encodes images beyond pixel level.
Experimental results demonstrate that the proposed RL-ASC is noise robust and could reconstruct visually pleasant and semantic consistent image.
arXiv Detail & Related papers (2022-08-08T12:29:55Z) - Explaining Classifiers by Constructing Familiar Concepts [2.7514191327409714]
We propose a decoder that transforms the incomprehensible representation of neurons into a representation that is more similar to the domain a human is familiar with.
An extension of ClaDec allows trading comprehensibility and fidelity.
We show that ClaDec tends to highlight more relevant input areas to classification though outcomes depend on architecture.
arXiv Detail & Related papers (2022-03-07T12:21:06Z) - Deep ensembles in bioimage segmentation [74.01883650587321]
In this work, we propose an ensemble of convolutional neural networks (CNNs)
In ensemble methods, many different models are trained and then used for classification, the ensemble aggregates the outputs of the single classifiers.
The proposed ensemble is implemented by combining different backbone networks using the DeepLabV3+ and HarDNet environment.
arXiv Detail & Related papers (2021-12-24T05:54:21Z) - Controlled Caption Generation for Images Through Adversarial Attacks [85.66266989600572]
We study adversarial examples for vision and language models, which typically adopt a Convolutional Neural Network (i.e., CNN) for image feature extraction and a Recurrent Neural Network (RNN) for caption generation.
In particular, we investigate attacks on the visual encoder's hidden layer that is fed to the subsequent recurrent network.
We propose a GAN-based algorithm for crafting adversarial examples for neural image captioning that mimics the internal representation of the CNN.
arXiv Detail & Related papers (2021-07-07T07:22:41Z) - Image Captioning using Deep Stacked LSTMs, Contextual Word Embeddings
and Data Augmentation [1.2183405753834562]
We propose to use Inception-ResNet Convolutional Neural Network as encoder to extract features from images.
We also use Hierarchical Context based Word Embeddings for word representations and a Deep Stacked Long Term Memory network as decoder.
We evaluate our proposed methods with two image captioning frameworks--Decoder and Soft Attention.
arXiv Detail & Related papers (2021-02-22T18:15:39Z) - Convolutional Neural Networks from Image Markers [62.997667081978825]
Feature Learning from Image Markers (FLIM) was recently proposed to estimate convolutional filters, with no backpropagation, from strokes drawn by a user on very few images.
This paper extends FLIM for fully connected layers and demonstrates it on different image classification problems.
The results show that FLIM-based convolutional neural networks can outperform the same architecture trained from scratch by backpropagation.
arXiv Detail & Related papers (2020-12-15T22:58:23Z) - Defending Adversarial Examples via DNN Bottleneck Reinforcement [20.08619981108837]
This paper presents a reinforcement scheme to alleviate the vulnerability of Deep Neural Networks (DNN) against adversarial attacks.
By reinforcing the former while maintaining the latter, any redundant information, be it adversarial or not, should be removed from the latent representation.
In order to reinforce the information bottleneck, we introduce the multi-scale low-pass objective and multi-scale high-frequency communication for better frequency steering in the network.
arXiv Detail & Related papers (2020-08-12T11:02:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.