Inherently Interpretable Multi-Label Classification Using Class-Specific
Counterfactuals
- URL: http://arxiv.org/abs/2303.00500v2
- Date: Tue, 8 Aug 2023 14:50:50 GMT
- Title: Inherently Interpretable Multi-Label Classification Using Class-Specific
Counterfactuals
- Authors: Susu Sun, Stefano Woerner, Andreas Maier, Lisa M. Koch, Christian F.
Baumgartner
- Abstract summary: Interpretability is essential for machine learning algorithms in high-stakes application fields such as medical image analysis.
We propose Attri-Net, an inherently interpretable model for multi-label classification.
We find that Attri-Net produces high-quality multi-label explanations consistent with clinical knowledge.
- Score: 9.485195366036292
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Interpretability is essential for machine learning algorithms in high-stakes
application fields such as medical image analysis. However, high-performing
black-box neural networks do not provide explanations for their predictions,
which can lead to mistrust and suboptimal human-ML collaboration. Post-hoc
explanation techniques, which are widely used in practice, have been shown to
suffer from severe conceptual problems. Furthermore, as we show in this paper,
current explanation techniques do not perform adequately in the multi-label
scenario, in which multiple medical findings may co-occur in a single image. We
propose Attri-Net, an inherently interpretable model for multi-label
classification. Attri-Net is a powerful classifier that provides transparent,
trustworthy, and human-understandable explanations. The model first generates
class-specific attribution maps based on counterfactuals to identify which
image regions correspond to certain medical findings. Then a simple logistic
regression classifier is used to make predictions based solely on these
attribution maps. We compare Attri-Net to five post-hoc explanation techniques
and one inherently interpretable classifier on three chest X-ray datasets. We
find that Attri-Net produces high-quality multi-label explanations consistent
with clinical knowledge and has comparable classification performance to
state-of-the-art classification models.
Related papers
- Attri-Net: A Globally and Locally Inherently Interpretable Model for Multi-Label Classification Using Class-Specific Counterfactuals [4.384272169863716]
Interpretability is crucial for machine learning algorithms in high-stakes medical applications.
Attri-Net is an inherently interpretable model for multi-label classification that provides local and global explanations.
arXiv Detail & Related papers (2024-06-08T13:52:02Z) - PM2: A New Prompting Multi-modal Model Paradigm for Few-shot Medical Image Classification [12.628447384868503]
We propose a new prompting multi-modal model paradigm on medical image classification based on multi-modal foundation models, called PM2.
Besides image modality, PM2 introduces another supplementary text input, known as prompt, to further describe corresponding image or concept classes.
Our PM2 significantly outperforms counterparts regardless of prompt schemes and achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-04-13T07:27:06Z) - Multi-task Explainable Skin Lesion Classification [54.76511683427566]
We propose a few-shot-based approach for skin lesions that generalizes well with few labelled data.
The proposed approach comprises a fusion of a segmentation network that acts as an attention module and classification network.
arXiv Detail & Related papers (2023-10-11T05:49:47Z) - Robust and Interpretable Medical Image Classifiers via Concept
Bottleneck Models [49.95603725998561]
We propose a new paradigm to build robust and interpretable medical image classifiers with natural language concepts.
Specifically, we first query clinical concepts from GPT-4, then transform latent image features into explicit concepts with a vision-language model.
arXiv Detail & Related papers (2023-10-04T21:57:09Z) - A Test Statistic Estimation-based Approach for Establishing
Self-interpretable CNN-based Binary Classifiers [7.424003880270276]
Post-hoc interpretability methods have the limitation that they can produce plausible but different interpretations.
The proposed method is self-interpretable, quantitative. Unlike the traditional post-hoc interpretability methods, the proposed method is self-interpretable, quantitative.
arXiv Detail & Related papers (2023-03-13T05:51:35Z) - Weakly-supervised segmentation using inherently-explainable
classification models and their application to brain tumour classification [0.46873264197900916]
This paper introduces three inherently-explainable classifiers to tackle both of these problems as one.
The models were employed for the task of multi-class brain tumour classification using two different datasets.
The obtained accuracy on a subset of tumour-only images outperformed the state-of-the-art glioma tumour grading binary classifiers with the best model achieving 98.7% accuracy.
arXiv Detail & Related papers (2022-06-10T14:44:05Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Theoretical Insights Into Multiclass Classification: A High-dimensional
Asymptotic View [82.80085730891126]
We provide the first modernally precise analysis of linear multiclass classification.
Our analysis reveals that the classification accuracy is highly distribution-dependent.
The insights gained may pave the way for a precise understanding of other classification algorithms.
arXiv Detail & Related papers (2020-11-16T05:17:29Z) - Improving Explainability of Image Classification in Scenarios with Class
Overlap: Application to COVID-19 and Pneumonia [7.372797734096181]
Trust in predictions made by machine learning models is increased if the model generalizes well on previously unseen samples.
We propose a method that enhances the explainability of image classifications through better localization by mitigating the model uncertainty induced by class overlap.
Our method is particularly promising in real-world class overlap scenarios, such as COVID-19 and pneumonia, where expertly labeled data for localization is not readily available.
arXiv Detail & Related papers (2020-08-06T20:47:36Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.