Explaining Deep Neural Networks using Unsupervised Clustering
- URL: http://arxiv.org/abs/2007.07477v2
- Date: Thu, 16 Jul 2020 00:50:15 GMT
- Title: Explaining Deep Neural Networks using Unsupervised Clustering
- Authors: Yu-han Liu and Sercan O. Arik
- Abstract summary: We propose a novel method to explain trained deep neural networks (DNNs) by distilling them into surrogate models using unsupervised clustering.
Our method can be applied flexibly to any subset of layers of a DNN architecture and can incorporate low-level and high-level information.
- Score: 12.639074798397619
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel method to explain trained deep neural networks (DNNs), by
distilling them into surrogate models using unsupervised clustering. Our method
can be applied flexibly to any subset of layers of a DNN architecture and can
incorporate low-level and high-level information. On image datasets given
pre-trained DNNs, we demonstrate the strength of our method in finding similar
training samples, and shedding light on the concepts the DNNs base their
decisions on. Via user studies, we show that our model can improve the user
trust in model's prediction.
Related papers
- Manipulating Feature Visualizations with Gradient Slingshots [54.31109240020007]
We introduce a novel method for manipulating Feature Visualization (FV) without significantly impacting the model's decision-making process.
We evaluate the effectiveness of our method on several neural network models and demonstrate its capabilities to hide the functionality of arbitrarily chosen neurons.
arXiv Detail & Related papers (2024-01-11T18:57:17Z) - A Gradient Boosting Approach for Training Convolutional and Deep Neural
Networks [0.0]
We introduce two procedures for training Convolutional Neural Networks (CNNs) and Deep Neural Network based on Gradient Boosting (GB)
The presented models show superior performance in terms of classification accuracy with respect to standard CNN and Deep-NN with the same architectures.
arXiv Detail & Related papers (2023-02-22T12:17:32Z) - Visualizing the Emergence of Intermediate Visual Patterns in DNNs [19.043540343193946]
This paper proposes a method to visualize the discrimination power of intermediate-layer visual patterns encoded by a DNN.
We visualize how the DNN gradually learns regional visual patterns in each intermediate layer during the training process.
This method also provides new insights into signal-processing behaviors of existing deep-learning techniques, such as adversarial attacks and knowledge distillation.
arXiv Detail & Related papers (2021-11-05T13:49:39Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Local Critic Training for Model-Parallel Learning of Deep Neural
Networks [94.69202357137452]
We propose a novel model-parallel learning method, called local critic training.
We show that the proposed approach successfully decouples the update process of the layer groups for both convolutional neural networks (CNNs) and recurrent neural networks (RNNs)
We also show that trained networks by the proposed method can be used for structural optimization.
arXiv Detail & Related papers (2021-02-03T09:30:45Z) - What Do Deep Nets Learn? Class-wise Patterns Revealed in the Input Space [88.37185513453758]
We propose a method to visualize and understand the class-wise knowledge learned by deep neural networks (DNNs) under different settings.
Our method searches for a single predictive pattern in the pixel space to represent the knowledge learned by the model for each class.
In the adversarial setting, we show that adversarially trained models tend to learn more simplified shape patterns.
arXiv Detail & Related papers (2021-01-18T06:38:41Z) - Model-Based Machine Learning for Communications [110.47840878388453]
We review existing strategies for combining model-based algorithms and machine learning from a high level perspective.
We focus on symbol detection, which is one of the fundamental tasks of communication receivers.
arXiv Detail & Related papers (2021-01-12T19:55:34Z) - Deep Neural Networks and Neuro-Fuzzy Networks for Intellectual Analysis
of Economic Systems [0.0]
We consider approaches for time series forecasting based on deep neural networks and neuro-fuzzy nets.
This paper presents also an overview of approaches for incorporating rule-based methodology into deep learning neural networks.
arXiv Detail & Related papers (2020-11-11T06:21:08Z) - Interpreting Graph Neural Networks for NLP With Differentiable Edge
Masking [63.49779304362376]
Graph neural networks (GNNs) have become a popular approach to integrating structural inductive biases into NLP models.
We introduce a post-hoc method for interpreting the predictions of GNNs which identifies unnecessary edges.
We show that we can drop a large proportion of edges without deteriorating the performance of the model.
arXiv Detail & Related papers (2020-10-01T17:51:19Z) - Causality-aware counterfactual confounding adjustment for feature
representations learned by deep models [14.554818659491644]
Causal modeling has been recognized as a potential solution to many challenging problems in machine learning (ML)
We describe how a recently proposed counterfactual approach can still be used to deconfound the feature representations learned by deep neural network (DNN) models.
arXiv Detail & Related papers (2020-04-20T17:37:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.