An Interpretable Neuron Embedding for Static Knowledge Distillation
- URL: http://arxiv.org/abs/2211.07647v1
- Date: Mon, 14 Nov 2022 03:26:10 GMT
- Title: An Interpretable Neuron Embedding for Static Knowledge Distillation
- Authors: Wei Han, Yangqiming Wang, Christian B\"ohm, Junming Shao
- Abstract summary: We propose a new interpretable neural network method, by embedding neurons into the semantic space.
The proposed semantic vector externalizes the latent knowledge to static knowledge, which is easy to exploit.
Empirical experiments of visualization show that semantic vectors describe neuron activation semantics well.
- Score: 7.644253344815002
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although deep neural networks have shown well-performance in various tasks,
the poor interpretability of the models is always criticized. In the paper, we
propose a new interpretable neural network method, by embedding neurons into
the semantic space to extract their intrinsic global semantics. In contrast to
previous methods that probe latent knowledge inside the model, the proposed
semantic vector externalizes the latent knowledge to static knowledge, which is
easy to exploit. Specifically, we assume that neurons with similar activation
are of similar semantic information. Afterwards, semantic vectors are optimized
by continuously aligning activation similarity and semantic vector similarity
during the training of the neural network. The visualization of semantic
vectors allows for a qualitative explanation of the neural network. Moreover,
we assess the static knowledge quantitatively by knowledge distillation tasks.
Empirical experiments of visualization show that semantic vectors describe
neuron activation semantics well. Without the sample-by-sample guidance from
the teacher model, static knowledge distillation exhibit comparable or even
superior performance with existing relation-based knowledge distillation
methods.
Related papers
- On the Value of Labeled Data and Symbolic Methods for Hidden Neuron Activation Analysis [1.55858752644861]
State of the art indicates that hidden node activations can, in some cases, be interpretable in a way that makes sense to humans.
We introduce a novel model-agnostic post-hoc Explainable AI method demonstrating that it provides meaningful interpretations.
arXiv Detail & Related papers (2024-04-21T07:57:45Z) - Improving Neural-based Classification with Logical Background Knowledge [0.0]
We propose a new formalism for supervised multi-label classification with propositional background knowledge.
We introduce a new neurosymbolic technique called semantic conditioning at inference.
We discuss its theoritical and practical advantages over two other popular neurosymbolic techniques.
arXiv Detail & Related papers (2024-02-20T14:01:26Z) - Hebbian Learning based Orthogonal Projection for Continual Learning of
Spiking Neural Networks [74.3099028063756]
We develop a new method with neuronal operations based on lateral connections and Hebbian learning.
We show that Hebbian and anti-Hebbian learning on recurrent lateral connections can effectively extract the principal subspace of neural activities.
Our method consistently solves for spiking neural networks with nearly zero forgetting.
arXiv Detail & Related papers (2024-02-19T09:29:37Z) - Manipulating Feature Visualizations with Gradient Slingshots [54.31109240020007]
We introduce a novel method for manipulating Feature Visualization (FV) without significantly impacting the model's decision-making process.
We evaluate the effectiveness of our method on several neural network models and demonstrate its capabilities to hide the functionality of arbitrarily chosen neurons.
arXiv Detail & Related papers (2024-01-11T18:57:17Z) - Understanding CNN Hidden Neuron Activations Using Structured Background
Knowledge and Deductive Reasoning [3.6223658572137825]
State of the art indicates that hidden node activations can, in some cases, be interpretable in a way that makes sense to humans.
We show that we can automatically attach meaningful labels from the background knowledge to individual neurons in the dense layer of a Convolutional Neural Network.
arXiv Detail & Related papers (2023-08-08T02:28:50Z) - Understanding Activation Patterns in Artificial Neural Networks by
Exploring Stochastic Processes [0.0]
We propose utilizing the framework of processes, which has been underutilized thus far.
We focus solely on activation frequency, leveraging neuroscience techniques used for real neuron spike trains.
We derive parameters describing activation patterns in each network, revealing consistent differences across architectures and training sets.
arXiv Detail & Related papers (2023-08-01T22:12:30Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Drop, Swap, and Generate: A Self-Supervised Approach for Generating
Neural Activity [33.06823702945747]
We introduce a novel unsupervised approach for learning disentangled representations of neural activity called Swap-VAE.
Our approach combines a generative modeling framework with an instance-specific alignment loss.
We show that it is possible to build representations that disentangle neural datasets along relevant latent dimensions linked to behavior.
arXiv Detail & Related papers (2021-11-03T16:39:43Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - Compositional Explanations of Neurons [52.71742655312625]
We describe a procedure for explaining neurons in deep representations by identifying compositional logical concepts.
We use this procedure to answer several questions on interpretability in models for vision and natural language processing.
arXiv Detail & Related papers (2020-06-24T20:37:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.