DProtoNet: Decoupling the inference module and the explanation module
enables neural networks to have better accuracy and interpretability
- URL: http://arxiv.org/abs/2210.08336v1
- Date: Sat, 15 Oct 2022 17:05:55 GMT
- Title: DProtoNet: Decoupling the inference module and the explanation module
enables neural networks to have better accuracy and interpretability
- Authors: Yitao Peng, Yihang Liu, Longzhen Yang, Lianghua He
- Abstract summary: In the previous method, by modifying the architecture of the neural network, the network simulates the human reasoning process.
We propose DProtoNet (Decoupling Prototypical network), it stores the decision basis of the neural network by using feature masks.
It decouples the neural network inference module from the interpretation module, and removes the specific architectural limitations of the interpretable network.
- Score: 5.333582981327497
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The interpretation of decisions made by neural networks is the focus of
recent research. In the previous method, by modifying the architecture of the
neural network, the network simulates the human reasoning process, that is, by
finding the decision elements to make decisions, so that the network has the
interpretability of the reasoning process. The specific interpretable
architecture will limit the fitting space of the network, resulting in a
decrease in the classification performance of the network, unstable
convergence, and general interpretability. We propose DProtoNet (Decoupling
Prototypical network), it stores the decision basis of the neural network by
using feature masks, and it uses Multiple Dynamic Masks (MDM) to explain the
decision basis for feature mask retention. It decouples the neural network
inference module from the interpretation module, and removes the specific
architectural limitations of the interpretable network, so that the
decision-making architecture of the network retains the original network
architecture as much as possible, making the neural network more expressive,
and greatly improving the interpretability. Classification performance and
interpretability of explanatory networks. We propose to replace the prototype
learning of a single image with the prototype learning of multiple images,
which makes the prototype robust, improves the convergence speed of network
training, and makes the accuracy of the network more stable during the learning
process. We test on multiple datasets, DProtoNet can improve the accuracy of
recent advanced interpretable network models by 5% to 10%, and its
classification performance is comparable to that of backbone networks without
interpretability. It also achieves the state of the art in interpretability
performance.
Related papers
- Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - Coding schemes in neural networks learning classification tasks [52.22978725954347]
We investigate fully-connected, wide neural networks learning classification tasks.
We show that the networks acquire strong, data-dependent features.
Surprisingly, the nature of the internal representations depends crucially on the neuronal nonlinearity.
arXiv Detail & Related papers (2024-06-24T14:50:05Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Semantic Loss Functions for Neuro-Symbolic Structured Prediction [74.18322585177832]
We discuss the semantic loss, which injects knowledge about such structure, defined symbolically, into training.
It is agnostic to the arrangement of the symbols, and depends only on the semantics expressed thereby.
It can be combined with both discriminative and generative neural models.
arXiv Detail & Related papers (2024-05-12T22:18:25Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - MDM:Visual Explanations for Neural Networks via Multiple Dynamic Mask [5.333582981327497]
We propose an algorithm Multiple Dynamic Mask(MDM), which is a general saliency graph query method with interpretability of the inference process.
For the MDM saliency map search algorithm, we experimentally compared the performance indicators of various saliency map search methods and the MDM with ResNet and DenseNet as the trained neural networks.
arXiv Detail & Related papers (2022-07-17T00:25:16Z) - Interpretable part-whole hierarchies and conceptual-semantic
relationships in neural networks [4.153804257347222]
We present Agglomerator, a framework capable of providing a representation of part-whole hierarchies from visual cues.
We evaluate our method on common datasets, such as SmallNORB, MNIST, FashionMNIST, CIFAR-10, and CIFAR-100.
arXiv Detail & Related papers (2022-03-07T10:56:13Z) - Modeling Structure with Undirected Neural Networks [20.506232306308977]
We propose undirected neural networks, a flexible framework for specifying computations that can be performed in any order.
We demonstrate the effectiveness of undirected neural architectures, both unstructured and structured, on a range of tasks.
arXiv Detail & Related papers (2022-02-08T10:06:51Z) - Creating Powerful and Interpretable Models withRegression Networks [2.2049183478692584]
We propose a novel architecture, Regression Networks, which combines the power of neural networks with the understandability of regression analysis.
We demonstrate that the models exceed the state-of-the-art performance of interpretable models on several benchmark datasets.
arXiv Detail & Related papers (2021-07-30T03:37:00Z) - Obtaining Faithful Interpretations from Compositional Neural Networks [72.41100663462191]
We evaluate the intermediate outputs of NMNs on NLVR2 and DROP datasets.
We find that the intermediate outputs differ from the expected output, illustrating that the network structure does not provide a faithful explanation of model behaviour.
arXiv Detail & Related papers (2020-05-02T06:50:35Z) - Inferring Convolutional Neural Networks' accuracies from their
architectural characterizations [0.0]
We study the relationships between a CNN's architecture and its performance.
We show that the attributes can be predictive of the networks' performance in two specific computer vision-based physics problems.
We use machine learning models to predict whether a network can perform better than a certain threshold accuracy before training.
arXiv Detail & Related papers (2020-01-07T16:41:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.