Can Physician Judgment Enhance Model Trustworthiness? A Case Study on
Predicting Pathological Lymph Nodes in Rectal Cancer
- URL: http://arxiv.org/abs/2312.09529v1
- Date: Fri, 15 Dec 2023 04:36:13 GMT
- Title: Can Physician Judgment Enhance Model Trustworthiness? A Case Study on
Predicting Pathological Lymph Nodes in Rectal Cancer
- Authors: Kazuma Kobayashi, Yasuyuki Takamizawa, Mototaka Miyake, Sono Ito, Lin
Gu, Tatsuya Nakatsuka, Yu Akagi, Tatsuya Harada, Yukihide Kanemitsu, Ryuji
Hamamoto
- Abstract summary: We employed a transformer to predict lymph node metastasis in rectal cancer using clinical data and magnetic resonance imaging.
We estimated the model's uncertainty using meta-level information like prediction probability variance and quantified agreement.
Our assessment of whether this agreement reduces uncertainty found no significant effect.
- Score: 35.293442328112036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Explainability is key to enhancing artificial intelligence's trustworthiness
in medicine. However, several issues remain concerning the actual benefit of
explainable models for clinical decision-making. Firstly, there is a lack of
consensus on an evaluation framework for quantitatively assessing the practical
benefits that effective explainability should provide to practitioners.
Secondly, physician-centered evaluations of explainability are limited.
Thirdly, the utility of built-in attention mechanisms in transformer-based
models as an explainability technique is unclear. We hypothesize that superior
attention maps should align with the information that physicians focus on,
potentially reducing prediction uncertainty and increasing model reliability.
We employed a multimodal transformer to predict lymph node metastasis in rectal
cancer using clinical data and magnetic resonance imaging, exploring how well
attention maps, visualized through a state-of-the-art technique, can achieve
agreement with physician understanding. We estimated the model's uncertainty
using meta-level information like prediction probability variance and
quantified agreement. Our assessment of whether this agreement reduces
uncertainty found no significant effect. In conclusion, this case study did not
confirm the anticipated benefit of attention maps in enhancing model
reliability. Superficial explanations could do more harm than good by
misleading physicians into relying on uncertain predictions, suggesting that
the current state of attention mechanisms in explainability should not be
overestimated. Identifying explainability mechanisms truly beneficial for
clinical decision-making remains essential.
Related papers
- Explainability of AI Uncertainty: Application to Multiple Sclerosis Lesion Segmentation on MRI [30.086986374579414]
This study introduces a novel framework to explain the potential sources of predictive uncertainty.
The proposed analysis shifts the focus from the uncertainty-error relationship towards relevant medical and engineering factors.
Our findings reveal that instance-wise uncertainty is strongly related to lesion size, shape, and cortical involvement.
arXiv Detail & Related papers (2025-04-07T08:09:27Z) - Reasoning-Enhanced Healthcare Predictions with Knowledge Graph Community Retrieval [61.70489848327436]
KARE is a novel framework that integrates knowledge graph (KG) community-level retrieval with large language models (LLMs) reasoning.
Extensive experiments demonstrate that KARE outperforms leading models by up to 10.8-15.0% on MIMIC-III and 12.6-12.7% on MIMIC-IV for mortality and readmission predictions.
arXiv Detail & Related papers (2024-10-06T18:46:28Z) - SepsisLab: Early Sepsis Prediction with Uncertainty Quantification and Active Sensing [67.8991481023825]
Sepsis is the leading cause of in-hospital mortality in the USA.
Existing predictive models are usually trained on high-quality data with few missing information.
For the potential high-risk patients with low confidence due to limited observations, we propose a robust active sensing algorithm.
arXiv Detail & Related papers (2024-07-24T04:47:36Z) - Assisting clinical practice with fuzzy probabilistic decision trees [2.0999441362198907]
We propose FPT, a novel method that combines probabilistic trees and fuzzy logic to assist clinical practice.
We show that FPT and its predictions can assist clinical practice in an intuitive manner, with the use of a user-friendly interface specifically designed for this purpose.
arXiv Detail & Related papers (2023-04-16T14:05:16Z) - Informing clinical assessment by contextualizing post-hoc explanations
of risk prediction models in type-2 diabetes [50.8044927215346]
We consider a comorbidity risk prediction scenario and focus on contexts regarding the patients clinical state.
We employ several state-of-the-art LLMs to present contexts around risk prediction model inferences and evaluate their acceptability.
Our paper is one of the first end-to-end analyses identifying the feasibility and benefits of contextual explanations in a real-world clinical use case.
arXiv Detail & Related papers (2023-02-11T18:07:11Z) - Towards Reliable Medical Image Segmentation by utilizing Evidential Calibrated Uncertainty [52.03490691733464]
We introduce DEviS, an easily implementable foundational model that seamlessly integrates into various medical image segmentation networks.
By leveraging subjective logic theory, we explicitly model probability and uncertainty for the problem of medical image segmentation.
DeviS incorporates an uncertainty-aware filtering module, which utilizes the metric of uncertainty-calibrated error to filter reliable data.
arXiv Detail & Related papers (2023-01-01T05:02:46Z) - Improving Trustworthiness of AI Disease Severity Rating in Medical
Imaging with Ordinal Conformal Prediction Sets [0.7734726150561088]
A lack of statistically rigorous uncertainty quantification is a significant factor undermining trust in AI results.
Recent developments in distribution-free uncertainty quantification present practical solutions for these issues.
We demonstrate a technique for forming ordinal prediction sets that are guaranteed to contain the correct stenosis severity.
arXiv Detail & Related papers (2022-07-05T18:01:20Z) - Benchmarking Heterogeneous Treatment Effect Models through the Lens of
Interpretability [82.29775890542967]
Estimating personalized effects of treatments is a complex, yet pervasive problem.
Recent developments in the machine learning literature on heterogeneous treatment effect estimation gave rise to many sophisticated, but opaque, tools.
We use post-hoc feature importance methods to identify features that influence the model's predictions.
arXiv Detail & Related papers (2022-06-16T17:59:05Z) - Uncertainty-Informed Deep Learning Models Enable High-Confidence
Predictions for Digital Histopathology [40.96261204117952]
We train models to identify lung adenocarcinoma vs. squamous cell carcinoma and show that high-confidence predictions outperform predictions without UQ.
We show that UQ thresholding remains reliable in the setting of domain shift, with accurate high-confidence predictions of adenocarcinoma vs. squamous cell carcinoma for out-of-distribution, non-lung cancer cohorts.
arXiv Detail & Related papers (2022-04-09T17:35:37Z) - Uncertainty aware and explainable diagnosis of retinal disease [0.0]
We perform uncertainty analysis of a deep learning model for diagnosis of four retinal diseases.
We show the features that a system used to make prediction while uncertainty awareness is the ability of a system to highlight when it is not sure about the decision.
arXiv Detail & Related papers (2021-01-26T23:37:30Z) - Bayesian Modelling in Practice: Using Uncertainty to Improve Trustworthiness in Medical Applications [2.446672595462589]
The Intensive Care Unit (ICU) is a hospital department where machine learning has the potential to provide valuable assistance in clinical decision making.
In practice, uncertain predictions should be presented to doctors with extra care in order to prevent potentially catastrophic treatment decisions.
We show how Bayesian modelling and the predictive uncertainty that it provides can be used to mitigate risk of misguided prediction.
arXiv Detail & Related papers (2019-06-20T13:51:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.