UbiQTree: Uncertainty Quantification in XAI with Tree Ensembles
- URL: http://arxiv.org/abs/2508.09639v1
- Date: Wed, 13 Aug 2025 09:20:33 GMT
- Title: UbiQTree: Uncertainty Quantification in XAI with Tree Ensembles
- Authors: Akshat Dubey, Aleksandar Anžel, Bahar İlgen, Georges Hattab,
- Abstract summary: We propose an approach for decomposing uncertainty in SHAP values into aleatoric, epistemic, and entanglement components.<n>We validate the method across three real-world use cases with descriptive statistical analyses.<n>This understanding can guide the development of robust decision-making processes and the refinement of models in high-stakes applications.
- Score: 42.37986459997699
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Explainable Artificial Intelligence (XAI) techniques, such as SHapley Additive exPlanations (SHAP), have become essential tools for interpreting complex ensemble tree-based models, especially in high-stakes domains such as healthcare analytics. However, SHAP values are usually treated as point estimates, which disregards the inherent and ubiquitous uncertainty in predictive models and data. This uncertainty has two primary sources: aleatoric and epistemic. The aleatoric uncertainty, which reflects the irreducible noise in the data. The epistemic uncertainty, which arises from a lack of data. In this work, we propose an approach for decomposing uncertainty in SHAP values into aleatoric, epistemic, and entanglement components. This approach integrates Dempster-Shafer evidence theory and hypothesis sampling via Dirichlet processes over tree ensembles. We validate the method across three real-world use cases with descriptive statistical analyses that provide insight into the nature of epistemic uncertainty embedded in SHAP explanations. The experimentations enable to provide more comprehensive understanding of the reliability and interpretability of SHAP-based attributions. This understanding can guide the development of robust decision-making processes and the refinement of models in high-stakes applications. Through our experiments with multiple datasets, we concluded that features with the highest SHAP values are not necessarily the most stable. This epistemic uncertainty can be reduced through better, more representative data and following appropriate or case-desired model development techniques. Tree-based models, especially bagging, facilitate the effective quantification of epistemic uncertainty.
Related papers
- SphUnc: Hyperspherical Uncertainty Decomposition and Causal Identification via Information Geometry [7.816699755198432]
We introduce SphUnc, a unified framework combining hyperspherical representation learning with structural causal modeling.<n>A structural causal model on spherical latents enables directed influence identification and interventional reasoning via sample-based simulation.<n> Empirical evaluations on social and affective benchmarks demonstrate improved accuracy, better calibration, and interpretable causal signals.
arXiv Detail & Related papers (2026-03-01T16:11:49Z) - Uncertainty Propagation in XAI: A Comparison of Analytical and Empirical Estimators [1.0855602842179624]
Understanding uncertainty in Explainable AI (XAI) is crucial for building trust.<n>This paper introduces a unified framework for quantifying and interpreting Uncertainty in XAI.<n>By using both analytical and empirical estimates of explanation variance, we provide a systematic means of assessing the impact uncertainty on explanations.
arXiv Detail & Related papers (2025-04-01T07:06:31Z) - Nonparametric Factor Analysis and Beyond [14.232694150264628]
We propose a general framework for identifying latent variables in the non-negligible settings.<n>We show that the generative model is identifiable up to certain submanifold indeterminacies even in the presence of non-negligible noise.<n>We have also developed corresponding estimation methods and validated them in various synthetic and real-world settings.
arXiv Detail & Related papers (2025-03-21T05:45:03Z) - Identifying Drivers of Predictive Aleatoric Uncertainty [2.5311562666866494]
We propose a straightforward approach to explain predictive aleatoric uncertainties.<n>We estimate uncertainty in regression as predictive variance by adapting a neural network with a Gaussian output distribution.<n>This approach can explain uncertainty influences more reliably than complex published approaches.
arXiv Detail & Related papers (2023-12-12T13:28:53Z) - Prototype-based Aleatoric Uncertainty Quantification for Cross-modal
Retrieval [139.21955930418815]
Cross-modal Retrieval methods build similarity relations between vision and language modalities by jointly learning a common representation space.
However, the predictions are often unreliable due to the Aleatoric uncertainty, which is induced by low-quality data, e.g., corrupt images, fast-paced videos, and non-detailed texts.
We propose a novel Prototype-based Aleatoric Uncertainty Quantification (PAU) framework to provide trustworthy predictions by quantifying the uncertainty arisen from the inherent data ambiguity.
arXiv Detail & Related papers (2023-09-29T09:41:19Z) - Measuring and Modeling Uncertainty Degree for Monocular Depth Estimation [50.920911532133154]
The intrinsic ill-posedness and ordinal-sensitive nature of monocular depth estimation (MDE) models pose major challenges to the estimation of uncertainty degree.
We propose to model the uncertainty of MDE models from the perspective of the inherent probability distributions.
By simply introducing additional training regularization terms, our model, with surprisingly simple formations and without requiring extra modules or multiple inferences, can provide uncertainty estimations with state-of-the-art reliability.
arXiv Detail & Related papers (2023-07-19T12:11:15Z) - An Epistemic and Aleatoric Decomposition of Arbitrariness to Constrain the Set of Good Models [7.620967781722717]
Recent research reveals that machine learning (ML) models are highly sensitive to minor changes in their training procedure.<n>We show that stability decomposes into epistemic and aleatoric components, capturing the consistency and confidence in prediction.<n>We propose a model selection procedure that includes epistemic and aleatoric criteria alongside existing accuracy and fairness criteria, and show that it successfully narrows down a large set of good models.
arXiv Detail & Related papers (2023-02-09T09:35:36Z) - Probabilities Are Not Enough: Formal Controller Synthesis for Stochastic
Dynamical Models with Epistemic Uncertainty [68.00748155945047]
Capturing uncertainty in models of complex dynamical systems is crucial to designing safe controllers.
Several approaches use formal abstractions to synthesize policies that satisfy temporal specifications related to safety and reachability.
Our contribution is a novel abstraction-based controller method for continuous-state models with noise, uncertain parameters, and external disturbances.
arXiv Detail & Related papers (2022-10-12T07:57:03Z) - When in Doubt: Neural Non-Parametric Uncertainty Quantification for
Epidemic Forecasting [70.54920804222031]
Most existing forecasting models disregard uncertainty quantification, resulting in mis-calibrated predictions.
Recent works in deep neural models for uncertainty-aware time-series forecasting also have several limitations.
We model the forecasting task as a probabilistic generative process and propose a functional neural process model called EPIFNP.
arXiv Detail & Related papers (2021-06-07T18:31:47Z) - The Hidden Uncertainty in a Neural Networks Activations [105.4223982696279]
The distribution of a neural network's latent representations has been successfully used to detect out-of-distribution (OOD) data.
This work investigates whether this distribution correlates with a model's epistemic uncertainty, thus indicating its ability to generalise to novel inputs.
arXiv Detail & Related papers (2020-12-05T17:30:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.