Concept-level Debugging of Part-Prototype Networks
- URL: http://arxiv.org/abs/2205.15769v1
- Date: Tue, 31 May 2022 13:18:51 GMT
- Title: Concept-level Debugging of Part-Prototype Networks
- Authors: Andrea Bontempelli, Stefano Teso, Fausto Giunchiglia, Andrea Passerini
- Abstract summary: Part-prototype Networks (ProtoPNets) are concept-based classifiers designed to achieve the same performance as black-box models without compromising transparency.
We propose a concept-level debugger for ProtoPNets in which a human supervisor supplies feedback in the form of what part-prototypes must be forgotten or kept.
- Score: 27.700043183428807
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Part-prototype Networks (ProtoPNets) are concept-based classifiers designed
to achieve the same performance as black-box models without compromising
transparency. ProtoPNets compute predictions based on similarity to
class-specific part-prototypes learned to recognize parts of training examples,
making it easy to faithfully determine what examples are responsible for any
target prediction and why. However, like other models, they are prone to
picking up confounds and shortcuts from the data, thus suffering from
compromised prediction accuracy and limited generalization. We propose
ProtoPDebug, an effective concept-level debugger for ProtoPNets in which a
human supervisor, guided by the model's explanations, supplies feedback in the
form of what part-prototypes must be forgotten or kept, and the model is
fine-tuned to align with this supervision. An extensive empirical evaluation on
synthetic and real-world data shows that ProtoPDebug outperforms
state-of-the-art debuggers for a fraction of the annotation cost.
Related papers
- Sparse Prototype Network for Explainable Pedestrian Behavior Prediction [60.80524827122901]
We present Sparse Prototype Network (SPN), an explainable method designed to simultaneously predict a pedestrian's future action, trajectory, and pose.
Regularized by mono-semanticity and clustering constraints, the prototypes learn consistent and human-understandable features.
arXiv Detail & Related papers (2024-10-16T03:33:40Z) - This actually looks like that: Proto-BagNets for local and global interpretability-by-design [5.037593461859481]
Interpretability is a key requirement for the use of machine learning models in high-stakes applications.
We introduce Proto-BagNets, an interpretable-by-design prototype-based model.
Proto-BagNet provides faithful, accurate, and clinically meaningful local and global explanations.
arXiv Detail & Related papers (2024-06-21T14:12:15Z) - Decomposing and Editing Predictions by Modeling Model Computation [75.37535202884463]
We introduce a task called component modeling.
The goal of component modeling is to decompose an ML model's prediction in terms of its components.
We present COAR, a scalable algorithm for estimating component attributions.
arXiv Detail & Related papers (2024-04-17T16:28:08Z) - MCPNet: An Interpretable Classifier via Multi-Level Concept Prototypes [24.28807025839685]
We argue that explanations lacking insights into the decision processes of low and mid-level features are neither fully faithful nor useful.
We propose a novel paradigm that learns and aligns multi-level concept prototype distributions for classification purposes via Class-aware Concept Distribution (CCD) loss.
arXiv Detail & Related papers (2024-04-13T11:13:56Z) - Uncertainty-Aware Explanations Through Probabilistic Self-Explainable Neural Networks [17.238290206236027]
Prototype-Based Self-Explainable Neural Networks offer deep, yet transparent-by-design architecture.
We introduce a probabilistic reformulation of PSENNs, called Prob-PSENN, which replaces point estimates for the prototypes with probability distributions over their values.
Our experiments demonstrate that Prob-PSENNs provide more meaningful and robust explanations than their non-probabilistic counterparts.
arXiv Detail & Related papers (2024-03-20T16:47:28Z) - ProtoP-OD: Explainable Object Detection with Prototypical Parts [0.0]
This paper introduces an extension to detection transformers that constructs prototypical local features and uses them in object detection.
The proposed extension consists of a bottleneck module, the prototype neck, that computes a discretized representation of prototype activations.
arXiv Detail & Related papers (2024-02-29T13:25:15Z) - ProtoSeg: Interpretable Semantic Segmentation with Prototypical Parts [12.959270094693254]
We introduce ProtoSeg, a novel model for interpretable semantic image segmentation.
To achieve accuracy comparable to baseline methods, we adapt the mechanism of prototypical parts.
We show that ProtoSeg discovers semantic concepts, in contrast to standard segmentation models.
arXiv Detail & Related papers (2023-01-28T19:14:32Z) - Pathologies of Pre-trained Language Models in Few-shot Fine-tuning [50.3686606679048]
We show that pre-trained language models with few examples show strong prediction bias across labels.
Although few-shot fine-tuning can mitigate the prediction bias, our analysis shows models gain performance improvement by capturing non-task-related features.
These observations alert that pursuing model performance with fewer examples may incur pathological prediction behavior.
arXiv Detail & Related papers (2022-04-17T15:55:18Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - Attentional Prototype Inference for Few-Shot Segmentation [128.45753577331422]
We propose attentional prototype inference (API), a probabilistic latent variable framework for few-shot segmentation.
We define a global latent variable to represent the prototype of each object category, which we model as a probabilistic distribution.
We conduct extensive experiments on four benchmarks, where our proposal obtains at least competitive and often better performance than state-of-the-art prototype-based methods.
arXiv Detail & Related papers (2021-05-14T06:58:44Z) - Prototypical Contrastive Learning of Unsupervised Representations [171.3046900127166]
Prototypical Contrastive Learning (PCL) is an unsupervised representation learning method.
PCL implicitly encodes semantic structures of the data into the learned embedding space.
PCL outperforms state-of-the-art instance-wise contrastive learning methods on multiple benchmarks.
arXiv Detail & Related papers (2020-05-11T09:53:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.