Explain by Evidence: An Explainable Memory-based Neural Network for
Question Answering
- URL: http://arxiv.org/abs/2011.03096v1
- Date: Thu, 5 Nov 2020 21:18:21 GMT
- Title: Explain by Evidence: An Explainable Memory-based Neural Network for
Question Answering
- Authors: Quan Tran, Nhan Dam, Tuan Lai, Franck Dernoncourt, Trung Le, Nham Le
and Dinh Phung
- Abstract summary: This paper proposes an explainable, evidence-based memory network architecture.
It learns to summarize the dataset and extract supporting evidences to make its decision.
Our model achieves state-of-the-art performance on two popular question answering datasets.
- Score: 41.73026155036886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Interpretability and explainability of deep neural networks are challenging
due to their scale, complexity, and the agreeable notions on which the
explaining process rests. Previous work, in particular, has focused on
representing internal components of neural networks through human-friendly
visuals and concepts. On the other hand, in real life, when making a decision,
human tends to rely on similar situations and/or associations in the past.
Hence arguably, a promising approach to make the model transparent is to design
it in a way such that the model explicitly connects the current sample with the
seen ones, and bases its decision on these samples. Grounded on that principle,
we propose in this paper an explainable, evidence-based memory network
architecture, which learns to summarize the dataset and extract supporting
evidences to make its decision. Our model achieves state-of-the-art performance
on two popular question answering datasets (i.e. TrecQA and WikiQA). Via
further analysis, we show that this model can reliably trace the errors it has
made in the validation step to the training instances that might have caused
these errors. We believe that this error-tracing capability provides
significant benefit in improving dataset quality in many applications.
Related papers
- NeuroInspect: Interpretable Neuron-based Debugging Framework through
Class-conditional Visualizations [28.552283701883766]
We present NeuroInspect, an interpretable neuron-based debug framework for deep learning (DL) models.
Our framework first pinpoints neurons responsible for mistakes in the network and then visualizes features embedded in the neurons to be human-interpretable.
We validate our framework by addressing false correlations and improving inferences for classes with the worst performance in real-world settings.
arXiv Detail & Related papers (2023-10-11T04:20:32Z) - Attributing Learned Concepts in Neural Networks to Training Data [5.930268338525991]
We find evidence for convergence, where removing the 10,000 top attributing images for a concept and retraining the model does not change the location of the concept in the network.
This suggests that the features that inform the development of a concept are spread in a more diffuse manner across its exemplars, implying robustness in concept formation.
arXiv Detail & Related papers (2023-10-04T20:26:59Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - LAP: An Attention-Based Module for Concept Based Self-Interpretation and
Knowledge Injection in Convolutional Neural Networks [2.8948274245812327]
We propose a new attention-based pooling layer, called Local Attention Pooling (LAP), that accomplishes self-interpretability.
LAP is easily pluggable into any convolutional neural network, even the already trained ones.
LAP offers more valid human-understandable and faithful-to-the-model interpretations than the commonly used white-box explainer methods.
arXiv Detail & Related papers (2022-01-27T21:10:20Z) - Explainable Adversarial Attacks in Deep Neural Networks Using Activation
Profiles [69.9674326582747]
This paper presents a visual framework to investigate neural network models subjected to adversarial examples.
We show how observing these elements can quickly pinpoint exploited areas in a model.
arXiv Detail & Related papers (2021-03-18T13:04:21Z) - Structural Causal Models Are (Solvable by) Credal Networks [70.45873402967297]
Causal inferences can be obtained by standard algorithms for the updating of credal nets.
This contribution should be regarded as a systematic approach to represent structural causal models by credal networks.
Experiments show that approximate algorithms for credal networks can immediately be used to do causal inference in real-size problems.
arXiv Detail & Related papers (2020-08-02T11:19:36Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.