iCapsNets: Towards Interpretable Capsule Networks for Text
Classification
- URL: http://arxiv.org/abs/2006.00075v1
- Date: Sat, 16 May 2020 04:11:44 GMT
- Title: iCapsNets: Towards Interpretable Capsule Networks for Text
Classification
- Authors: Zhengyang Wang, Xia Hu, Shuiwang Ji
- Abstract summary: Traditional machine learning methods are easy to interpret but have low accuracies.
We propose interpretable capsule networks (iCapsNets) to bridge this gap.
iCapsNets can be interpreted both locally and globally.
- Score: 95.31786902390438
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many text classification applications require models with satisfying
performance as well as good interpretability. Traditional machine learning
methods are easy to interpret but have low accuracies. The development of deep
learning models boosts the performance significantly. However, deep learning
models are typically hard to interpret. In this work, we propose interpretable
capsule networks (iCapsNets) to bridge this gap. iCapsNets use capsules to
model semantic meanings and explore novel methods to increase interpretability.
The design of iCapsNets is consistent with human intuition and enables it to
produce human-understandable interpretation results. Notably, iCapsNets can be
interpreted both locally and globally. In terms of local interpretability,
iCapsNets offer a simple yet effective method to explain the predictions for
each data sample. On the other hand, iCapsNets explore a novel way to explain
the model's general behavior, achieving global interpretability. Experimental
studies show that our iCapsNets yield meaningful local and global
interpretation results, without suffering from significant performance loss
compared to non-interpretable methods.
Related papers
- Sparse Prototype Network for Explainable Pedestrian Behavior Prediction [60.80524827122901]
We present Sparse Prototype Network (SPN), an explainable method designed to simultaneously predict a pedestrian's future action, trajectory, and pose.
Regularized by mono-semanticity and clustering constraints, the prototypes learn consistent and human-understandable features.
arXiv Detail & Related papers (2024-10-16T03:33:40Z) - Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - This actually looks like that: Proto-BagNets for local and global interpretability-by-design [5.037593461859481]
Interpretability is a key requirement for the use of machine learning models in high-stakes applications.
We introduce Proto-BagNets, an interpretable-by-design prototype-based model.
Proto-BagNet provides faithful, accurate, and clinically meaningful local and global explanations.
arXiv Detail & Related papers (2024-06-21T14:12:15Z) - Seeing in Words: Learning to Classify through Language Bottlenecks [59.97827889540685]
Humans can explain their predictions using succinct and intuitive descriptions.
We show that a vision model whose feature representations are text can effectively classify ImageNet images.
arXiv Detail & Related papers (2023-06-29T00:24:42Z) - Towards the Characterization of Representations Learned via
Capsule-based Network Architectures [0.0]
Capsule Networks (CapsNets) have been re-introduced as a more compact and interpretable alternative to standard deep neural networks.
Here, we conduct a systematic and principled study towards assessing the interpretability of these types of networks.
Our analysis in the MNIST, SVHN, PASCAL-part and CelebA datasets suggest that the representations encoded in CapsNets might not be as disentangled nor strictly related to parts-whole relationships as is commonly stated in the literature.
arXiv Detail & Related papers (2023-05-09T11:20:11Z) - Language Model Pre-Training with Sparse Latent Typing [66.75786739499604]
We propose a new pre-training objective, Sparse Latent Typing, which enables the model to sparsely extract sentence-level keywords with diverse latent types.
Experimental results show that our model is able to learn interpretable latent type categories in a self-supervised manner without using any external knowledge.
arXiv Detail & Related papers (2022-10-23T00:37:08Z) - Parallel Capsule Networks for Classification of White Blood Cells [1.5749416770494706]
Capsule Networks (CapsNets) is a machine learning architecture proposed to overcome some of the shortcomings of convolutional neural networks (CNNs)
We present a new architecture, parallel CapsNets, which exploits the concept of branching the network to isolate certain capsules.
arXiv Detail & Related papers (2021-08-05T14:30:44Z) - Leveraging Sparse Linear Layers for Debuggable Deep Networks [86.94586860037049]
We show how fitting sparse linear models over learned deep feature representations can lead to more debuggable neural networks.
The resulting sparse explanations can help to identify spurious correlations, explain misclassifications, and diagnose model biases in vision and language tasks.
arXiv Detail & Related papers (2021-05-11T08:15:25Z) - Pushing the Limits of Capsule Networks [1.8231854497751137]
Convolutional neural networks do not explicitly maintain a representation of the locations of the features relative to each other.
A team at Google Brain recently made news with an attempt to fix this problem: Capsule Networks.
We want to stress test CapsNet in various incremental ways to better understand their performance and expressiveness.
arXiv Detail & Related papers (2021-03-15T00:30:34Z) - Interpretable Graph Capsule Networks for Object Recognition [17.62514568986647]
We propose interpretable Graph Capsule Networks (GraCapsNets), where we replace the routing part with a multi-head attention-based Graph Pooling approach.
GraCapsNets achieve better classification performance with fewer parameters and better adversarial robustness, when compared to CapsNets.
arXiv Detail & Related papers (2020-12-03T03:18:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.