ProtoryNet - Interpretable Text Classification Via Prototype
Trajectories
- URL: http://arxiv.org/abs/2007.01777v5
- Date: Mon, 6 Nov 2023 16:33:48 GMT
- Title: ProtoryNet - Interpretable Text Classification Via Prototype
Trajectories
- Authors: Dat Hong, Tong Wang, Stephen S. Baek
- Abstract summary: We propose a novel interpretable deep neural network for text classification, called ProtoryNet.
ProtoryNet makes a prediction by finding the most similar prototype for each sentence in a text sequence.
After prototype pruning, the resulting ProtoryNet models only need less than or around 20 prototypes for all datasets.
- Score: 4.768286204382179
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a novel interpretable deep neural network for text classification,
called ProtoryNet, based on a new concept of prototype trajectories. Motivated
by the prototype theory in modern linguistics, ProtoryNet makes a prediction by
finding the most similar prototype for each sentence in a text sequence and
feeding an RNN backbone with the proximity of each sentence to the
corresponding active prototype. The RNN backbone then captures the temporal
pattern of the prototypes, which we refer to as prototype trajectories.
Prototype trajectories enable intuitive and fine-grained interpretation of the
reasoning process of the RNN model, in resemblance to how humans analyze texts.
We also design a prototype pruning procedure to reduce the total number of
prototypes used by the model for better interpretability. Experiments on
multiple public data sets show that ProtoryNet is more accurate than the
baseline prototype-based deep neural net and reduces the performance gap
compared to state-of-the-art black-box models. In addition, after prototype
pruning, the resulting ProtoryNet models only need less than or around 20
prototypes for all datasets, which significantly benefits interpretability.
Furthermore, we report a survey result indicating that human users find
ProtoryNet more intuitive and easier to understand than other prototype-based
methods.
Related papers
- Advancing Interpretability in Text Classification through Prototype Learning [1.9526476410335776]
ProtoLens is a prototype-based model that provides fine-grained, sub-sentence level interpretability for text classification.
ProtoLens uses a Prototype-aware Span Extraction module to identify relevant text spans.
ProtoLens provides interpretable predictions while maintaining competitive accuracy.
arXiv Detail & Related papers (2024-10-23T03:53:46Z) - Sparse Prototype Network for Explainable Pedestrian Behavior Prediction [60.80524827122901]
We present Sparse Prototype Network (SPN), an explainable method designed to simultaneously predict a pedestrian's future action, trajectory, and pose.
Regularized by mono-semanticity and clustering constraints, the prototypes learn consistent and human-understandable features.
arXiv Detail & Related papers (2024-10-16T03:33:40Z) - This actually looks like that: Proto-BagNets for local and global interpretability-by-design [5.037593461859481]
Interpretability is a key requirement for the use of machine learning models in high-stakes applications.
We introduce Proto-BagNets, an interpretable-by-design prototype-based model.
Proto-BagNet provides faithful, accurate, and clinically meaningful local and global explanations.
arXiv Detail & Related papers (2024-06-21T14:12:15Z) - Query-guided Prototype Evolution Network for Few-Shot Segmentation [85.75516116674771]
We present a new method that integrates query features into the generation process of foreground and background prototypes.
Experimental results on the PASCAL-$5i$ and mirroring-$20i$ datasets attest to the substantial enhancements achieved by QPENet.
arXiv Detail & Related papers (2024-03-11T07:50:40Z) - MProto: Multi-Prototype Network with Denoised Optimal Transport for
Distantly Supervised Named Entity Recognition [75.87566793111066]
We propose a noise-robust prototype network named MProto for the DS-NER task.
MProto represents each entity type with multiple prototypes to characterize the intra-class variance.
To mitigate the noise from incomplete labeling, we propose a novel denoised optimal transport (DOT) algorithm.
arXiv Detail & Related papers (2023-10-12T13:02:34Z) - ProtoDiff: Learning to Learn Prototypical Networks by Task-Guided
Diffusion [44.805452233966534]
Prototype-based meta-learning has emerged as a powerful technique for addressing few-shot learning challenges.
We introduce ProtoDiff, a framework that gradually generates task-specific prototypes from random noise.
We conduct thorough ablation studies to demonstrate its ability to accurately capture the underlying prototype distribution.
arXiv Detail & Related papers (2023-06-26T15:26:24Z) - Rethinking Semantic Segmentation: A Prototype View [126.59244185849838]
We present a nonparametric semantic segmentation model based on non-learnable prototypes.
Our framework yields compelling results over several datasets.
We expect this work will provoke a rethink of the current de facto semantic segmentation model design.
arXiv Detail & Related papers (2022-03-28T21:15:32Z) - Deformable ProtoPNet: An Interpretable Image Classifier Using Deformable Prototypes [7.8515366468594765]
We present a deformable part network (Deformable ProtoPNet) that integrates the power of deep learning and the interpretability of case-based reasoning.
This model classifies input images by comparing them with prototypes learned during training, yielding explanations in the form of "this looks like that"
arXiv Detail & Related papers (2021-11-29T22:38:13Z) - Prototype-based interpretation of the functionality of neurons in
winner-take-all neural networks [1.418033127602866]
Prototype-based learning (PbL) using a winner-take-all (WTA) network based on minimum Euclidean distance (ED-WTA) is an intuitive approach to multiclass classification.
We propose a novel training algorithm for the $pm$ED-WTA network, which cleverly switches between updating the positive and negative prototypes.
We show that the proposed $pm$ED-WTA method constructs highly interpretable prototypes that can be successfully used for detecting and adversarial examples.
arXiv Detail & Related papers (2020-08-20T03:15:37Z) - Learning Sparse Prototypes for Text Generation [120.38555855991562]
Prototype-driven text generation is inefficient at test time as a result of needing to store and index the entire training corpus.
We propose a novel generative model that automatically learns a sparse prototype support set that achieves strong language modeling performance.
In experiments, our model outperforms previous prototype-driven language models while achieving up to a 1000x memory reduction.
arXiv Detail & Related papers (2020-06-29T19:41:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.