Investigating Poor Performance Regions of Black Boxes: LIME-based
Exploration in Sepsis Detection
- URL: http://arxiv.org/abs/2306.12507v1
- Date: Wed, 21 Jun 2023 18:36:15 GMT
- Title: Investigating Poor Performance Regions of Black Boxes: LIME-based
Exploration in Sepsis Detection
- Authors: Mozhgan Salimiparsa, Surajsinh Parmar, San Lee, Choongmin Kim,
Yonghwan Kim, Jang Yong Kim
- Abstract summary: This paper proposes leveraging Local Interpretable Model-Agnostic Explanations (LIME) to provide interpretable descriptions of black box classification models in sepsis detection.
By analyzing misclassified instances, significant features contributing to suboptimal performance are identified.
- Score: 0.5872014229110214
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Interpreting machine learning models remains a challenge, hindering their
adoption in clinical settings. This paper proposes leveraging Local
Interpretable Model-Agnostic Explanations (LIME) to provide interpretable
descriptions of black box classification models in high-stakes sepsis
detection. By analyzing misclassified instances, significant features
contributing to suboptimal performance are identified. The analysis reveals
regions where the classifier performs poorly, allowing the calculation of error
rates within these regions. This knowledge is crucial for cautious
decision-making in sepsis detection and other critical applications. The
proposed approach is demonstrated using the eICU dataset, effectively
identifying and visualizing regions where the classifier underperforms. By
enhancing interpretability, our method promotes the adoption of machine
learning models in clinical practice, empowering informed decision-making and
mitigating risks in critical scenarios.
Related papers
- Interpretability of Uncertainty: Exploring Cortical Lesion Segmentation in Multiple Sclerosis [33.91263917157504]
Uncertainty quantification (UQ) has become critical for evaluating the reliability of artificial intelligence systems.
This study addresses the interpretability of instance-wise uncertainty values in deep learning models for focal lesion segmentation in magnetic resonance imaging.
arXiv Detail & Related papers (2024-07-08T09:13:30Z) - Multi-task Explainable Skin Lesion Classification [54.76511683427566]
We propose a few-shot-based approach for skin lesions that generalizes well with few labelled data.
The proposed approach comprises a fusion of a segmentation network that acts as an attention module and classification network.
arXiv Detail & Related papers (2023-10-11T05:49:47Z) - SHAMSUL: Systematic Holistic Analysis to investigate Medical
Significance Utilizing Local interpretability methods in deep learning for
chest radiography pathology prediction [1.0138723409205497]
The study delves into the application of four well-established interpretability methods: Local Interpretable Model-agnostic Explanations (LIME), Shapley Additive exPlanations (SHAP), Gradient-weighted Class Activation Mapping (Grad-CAM) and Layer-wise Relevance Propagation (LRP)
Our analysis encompasses both single-label and multi-label predictions, providing a comprehensive and unbiased assessment through quantitative and qualitative investigations, which are compared against human expert annotation.
arXiv Detail & Related papers (2023-07-16T11:10:35Z) - Deployment of Image Analysis Algorithms under Prevalence Shifts [6.373765910269204]
Domain gaps are among the most relevant roadblocks in the clinical translation of machine learning (ML)-based solutions for medical image analysis.
We propose a workflow for prevalence-aware image classification that uses estimated deployment prevalences to adjust a trained classifier to a new environment.
arXiv Detail & Related papers (2023-03-22T13:16:37Z) - Plugin estimators for selective classification with out-of-distribution
detection [67.28226919253214]
Real-world classifiers can benefit from abstaining from predicting on samples where they have low confidence.
These settings have been the subject of extensive but disjoint study in the selective classification (SC) and out-of-distribution (OOD) detection literature.
Recent work on selective classification with OOD detection has argued for the unified study of these problems.
We propose new plugin estimators for SCOD that are theoretically grounded, effective, and generalise existing approaches.
arXiv Detail & Related papers (2023-01-29T07:45:17Z) - Benchmarking Heterogeneous Treatment Effect Models through the Lens of
Interpretability [82.29775890542967]
Estimating personalized effects of treatments is a complex, yet pervasive problem.
Recent developments in the machine learning literature on heterogeneous treatment effect estimation gave rise to many sophisticated, but opaque, tools.
We use post-hoc feature importance methods to identify features that influence the model's predictions.
arXiv Detail & Related papers (2022-06-16T17:59:05Z) - Improving Classification Model Performance on Chest X-Rays through Lung
Segmentation [63.45024974079371]
We propose a deep learning approach to enhance abnormal chest x-ray (CXR) identification performance through segmentations.
Our approach is designed in a cascaded manner and incorporates two modules: a deep neural network with criss-cross attention modules (XLSor) for localizing lung region in CXR images and a CXR classification model with a backbone of a self-supervised momentum contrast (MoCo) model pre-trained on large-scale CXR data sets.
arXiv Detail & Related papers (2022-02-22T15:24:06Z) - Spatial machine-learning model diagnostics: a model-agnostic
distance-based approach [91.62936410696409]
This contribution proposes spatial prediction error profiles (SPEPs) and spatial variable importance profiles (SVIPs) as novel model-agnostic assessment and interpretation tools.
The SPEPs and SVIPs of geostatistical methods, linear models, random forest, and hybrid algorithms show striking differences and also relevant similarities.
The novel diagnostic tools enrich the toolkit of spatial data science, and may improve ML model interpretation, selection, and design.
arXiv Detail & Related papers (2021-11-13T01:50:36Z) - MCDAL: Maximum Classifier Discrepancy for Active Learning [74.73133545019877]
Recent state-of-the-art active learning methods have mostly leveraged Generative Adversarial Networks (GAN) for sample acquisition.
We propose in this paper a novel active learning framework that we call Maximum Discrepancy for Active Learning (MCDAL)
In particular, we utilize two auxiliary classification layers that learn tighter decision boundaries by maximizing the discrepancies among them.
arXiv Detail & Related papers (2021-07-23T06:57:08Z) - Quantifying Explainability in NLP and Analyzing Algorithms for
Performance-Explainability Tradeoff [0.0]
We explore the current art of explainability and interpretability within a case study in clinical text classification.
We demonstrate various visualization techniques for fully interpretable methods as well as model-agnostic post hoc attributions.
We introduce a framework through which practitioners and researchers can assess the frontier between a model's predictive performance and the quality of its available explanations.
arXiv Detail & Related papers (2021-07-12T19:07:24Z) - An Efficient Framework for Automated Screening of Clinically Significant
Macular Edema [0.41998444721319206]
The present study proposes a new approach to automated screening of Clinically Significant Macular Edema (CSME)
The proposed approach combines a pre-trained deep neural network with meta-heuristic feature selection.
A feature space over-sampling technique is being used to overcome the effects of skewed datasets.
arXiv Detail & Related papers (2020-01-20T07:34:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.