DeepSeer: Interactive RNN Explanation and Debugging via State
Abstraction
- URL: http://arxiv.org/abs/2303.01576v1
- Date: Thu, 2 Mar 2023 21:08:17 GMT
- Title: DeepSeer: Interactive RNN Explanation and Debugging via State
Abstraction
- Authors: Zhijie Wang, Yuheng Huang, Da Song, Lei Ma, Tianyi Zhang
- Abstract summary: Recurrent Neural Networks (RNNs) have been widely used in Natural Language Processing (NLP) tasks.
DeepSeer is an interactive system that provides both global and local explanations of RNN behavior.
- Score: 10.110976560799612
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recurrent Neural Networks (RNNs) have been widely used in Natural Language
Processing (NLP) tasks given its superior performance on processing sequential
data. However, it is challenging to interpret and debug RNNs due to the
inherent complexity and the lack of transparency of RNNs. While many
explainable AI (XAI) techniques have been proposed for RNNs, most of them only
support local explanations rather than global explanations. In this paper, we
present DeepSeer, an interactive system that provides both global and local
explanations of RNN behavior in multiple tightly-coordinated views for model
understanding and debugging. The core of DeepSeer is a state abstraction method
that bundles semantically similar hidden states in an RNN model and abstracts
the model as a finite state machine. Users can explore the global model
behavior by inspecting text patterns associated with each state and the
transitions between states. Users can also dive into individual predictions by
inspecting the state trace and intermediate prediction results of a given
input. A between-subjects user study with 28 participants shows that, compared
with a popular XAI technique, LIME, participants using DeepSeer made a deeper
and more comprehensive assessment of RNN model behavior, identified the root
causes of incorrect predictions more accurately, and came up with more
actionable plans to improve the model performance.
Related papers
- Use of Parallel Explanatory Models to Enhance Transparency of Neural Network Configurations for Cell Degradation Detection [18.214293024118145]
We build a parallel model to illuminate and understand the internal operation of neural networks.
We show how each layer of the RNN transforms the input distributions to increase detection accuracy.
At the same time we also discover a side effect acting to limit the improvement in accuracy.
arXiv Detail & Related papers (2024-04-17T12:22:54Z) - Towards Better Explanations for Object Detection [0.0]
This paper proposes a method to explain the decision for any object detection model called D-CLOSE.
We performed tests on the MS-COCO dataset with the YOLOX model, which shows that our method outperforms D-RISE.
arXiv Detail & Related papers (2023-06-05T09:52:05Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Batch-Ensemble Stochastic Neural Networks for Out-of-Distribution
Detection [55.028065567756066]
Out-of-distribution (OOD) detection has recently received much attention from the machine learning community due to its importance in deploying machine learning models in real-world applications.
In this paper we propose an uncertainty quantification approach by modelling the distribution of features.
We incorporate an efficient ensemble mechanism, namely batch-ensemble, to construct the batch-ensemble neural networks (BE-SNNs) and overcome the feature collapse problem.
We show that BE-SNNs yield superior performance on several OOD benchmarks, such as the Two-Moons dataset, the FashionMNIST vs MNIST dataset, FashionM
arXiv Detail & Related papers (2022-06-26T16:00:22Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Extracting Finite Automata from RNNs Using State Merging [1.8072051868187933]
We propose a new method for extracting finite automata from RNNs inspired by the state merging paradigm from grammatical inference.
We demonstrate the effectiveness of our method on the Tomita languages benchmark, where we find that it is able to extract faithful automata from RNNs trained on all languages in the benchmark.
arXiv Detail & Related papers (2022-01-28T23:03:25Z) - Self-Ensembling GAN for Cross-Domain Semantic Segmentation [107.27377745720243]
This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
arXiv Detail & Related papers (2021-12-15T09:50:25Z) - i-Algebra: Towards Interactive Interpretability of Deep Neural Networks [41.13047686374529]
We present i-Algebra, a first-of-its-kind interactive framework for interpreting deep neural networks (DNNs)
At its core is a library of atomic, composable operators, which explain model behaviors at varying input granularity, during different inference stages, and from distinct interpretation perspectives.
We conduct user studies in a set of representative analysis tasks, including inspecting adversarial inputs, resolving model inconsistency, and cleansing contaminated data, all demonstrating its promising usability.
arXiv Detail & Related papers (2021-01-22T19:22:57Z) - Explaining and Improving Model Behavior with k Nearest Neighbor
Representations [107.24850861390196]
We propose using k nearest neighbor representations to identify training examples responsible for a model's predictions.
We show that kNN representations are effective at uncovering learned spurious associations.
Our results indicate that the kNN approach makes the finetuned model more robust to adversarial inputs.
arXiv Detail & Related papers (2020-10-18T16:55:25Z) - Interpreting Graph Neural Networks for NLP With Differentiable Edge
Masking [63.49779304362376]
Graph neural networks (GNNs) have become a popular approach to integrating structural inductive biases into NLP models.
We introduce a post-hoc method for interpreting the predictions of GNNs which identifies unnecessary edges.
We show that we can drop a large proportion of edges without deteriorating the performance of the model.
arXiv Detail & Related papers (2020-10-01T17:51:19Z) - An Adversarial Approach for Explaining the Predictions of Deep Neural
Networks [9.645196221785694]
We present a novel algorithm for explaining the predictions of a deep neural network (DNN) using adversarial machine learning.
Our approach identifies the relative importance of input features in relation to the predictions based on the behavior of an adversarial attack on the DNN.
Our analysis enables us to produce consistent and efficient explanations.
arXiv Detail & Related papers (2020-05-20T18:06:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.