Quantifying Adversarial Uncertainty in Evidential Deep Learning using Conflict Resolution
- URL: http://arxiv.org/abs/2506.05937v1
- Date: Fri, 06 Jun 2025 10:06:23 GMT
- Title: Quantifying Adversarial Uncertainty in Evidential Deep Learning using Conflict Resolution
- Authors: Charmaine Barker, Daniel Bethell, Simos Gerasimou,
- Abstract summary: Conflict-aware Evidential Deep Learning (C-EDL) is a lightweight post-hoc uncertainty quantification approach.<n>C-EDL generates diverse, task-preserving transformations per input and quantifies disagreement to calibrate uncertainty estimates.<n>Our experimental evaluation shows that C-EDL significantly outperforms state-of-the-art EDL variants and competitive baselines.
- Score: 2.321323878201932
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Reliability of deep learning models is critical for deployment in high-stakes applications, where out-of-distribution or adversarial inputs may lead to detrimental outcomes. Evidential Deep Learning, an efficient paradigm for uncertainty quantification, models predictions as Dirichlet distributions of a single forward pass. However, EDL is particularly vulnerable to adversarially perturbed inputs, making overconfident errors. Conflict-aware Evidential Deep Learning (C-EDL) is a lightweight post-hoc uncertainty quantification approach that mitigates these issues, enhancing adversarial and OOD robustness without retraining. C-EDL generates diverse, task-preserving transformations per input and quantifies representational disagreement to calibrate uncertainty estimates when needed. C-EDL's conflict-aware prediction adjustment improves detection of OOD and adversarial inputs, maintaining high in-distribution accuracy and low computational overhead. Our experimental evaluation shows that C-EDL significantly outperforms state-of-the-art EDL variants and competitive baselines, achieving substantial reductions in coverage for OOD data (up to 55%) and adversarial data (up to 90%), across a range of datasets, attack types, and uncertainty metrics.
Related papers
- Robust Molecular Property Prediction via Densifying Scarce Labeled Data [51.55434084913129]
In drug discovery, compounds most critical for advancing research often lie beyond the training set.<n>We propose a novel meta-learning-based approach that leverages unlabeled data to interpolate between in-distribution (ID) and out-of-distribution (OOD) data.<n>We demonstrate significant performance gains on challenging real-world datasets.
arXiv Detail & Related papers (2025-06-13T15:27:40Z) - Towards Harmonized Uncertainty Estimation for Large Language Models [22.58034272573749]
It is essential to quantify the reliability of their generations through uncertainty estimation.<n>We propose CUE (Corrector for Uncertainty Estimation): A straightforward yet effective method that employs a lightweight model trained on data aligned with the target LLM's performance to adjust uncertainty scores.
arXiv Detail & Related papers (2025-05-25T10:17:57Z) - Calibrating LLMs with Information-Theoretic Evidential Deep Learning [22.263501423275784]
Fine-tuned large language models (LLMs) often exhibit overconfidence, particularly when trained on small datasets.<n>Evidential Deep Learning (EDL), an uncertainty-aware approach, enables uncertainty estimation in a single forward pass.<n>We propose regularizing EDL by incorporating an information bottleneck (IB)
arXiv Detail & Related papers (2025-02-10T11:00:24Z) - Revisiting Essential and Nonessential Settings of Evidential Deep Learning [70.82728812001807]
Evidential Deep Learning (EDL) is an emerging method for uncertainty estimation.
We propose Re-EDL, a simplified yet more effective variant of EDL.
arXiv Detail & Related papers (2024-10-01T04:27:07Z) - Towards Robust Uncertainty-Aware Incomplete Multi-View Classification [11.617211995206018]
We propose the Alternating Progressive Learning Network (APLN) to enhance EDL-based methods in incomplete MVC scenarios.
APLN mitigates bias from corrupted observed data by first applying coarse imputation, followed by mapping the data to a latent space.
We also introduce a conflict-aware Dempster-Shafer combination rule (DSCR) to better handle conflicting evidence.
arXiv Detail & Related papers (2024-09-10T07:18:57Z) - A Comprehensive Survey on Evidential Deep Learning and Its Applications [64.83473301188138]
Evidential Deep Learning (EDL) provides reliable uncertainty estimation with minimal additional computation in a single forward pass.
We first delve into the theoretical foundation of EDL, the subjective logic theory, and discuss its distinctions from other uncertainty estimation frameworks.
We elaborate on its extensive applications across various machine learning paradigms and downstream tasks.
arXiv Detail & Related papers (2024-09-07T05:55:06Z) - Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling [69.83976050879318]
In large language models (LLMs), identifying sources of uncertainty is an important step toward improving reliability, trustworthiness, and interpretability.
In this paper, we introduce an uncertainty decomposition framework for LLMs, called input clarification ensembling.
Our approach generates a set of clarifications for the input, feeds them into an LLM, and ensembles the corresponding predictions.
arXiv Detail & Related papers (2023-11-15T05:58:35Z) - Knowledge from Uncertainty in Evidential Deep Learning [10.751990848772028]
This work reveals an evidential signal that emerges from the uncertainty value in Evidential Deep Learning (EDL)
EDL is one example of a class of uncertainty-aware deep learning approaches designed to provide confidence about the current test sample.
arXiv Detail & Related papers (2023-10-19T11:41:52Z) - Uncertainty Estimation by Fisher Information-based Evidential Deep
Learning [61.94125052118442]
Uncertainty estimation is a key factor that makes deep learning reliable in practical applications.
We propose a novel method, Fisher Information-based Evidential Deep Learning ($mathcalI$-EDL)
In particular, we introduce Fisher Information Matrix (FIM) to measure the informativeness of evidence carried by each sample, according to which we can dynamically reweight the objective loss terms to make the network more focused on the representation learning of uncertain classes.
arXiv Detail & Related papers (2023-03-03T16:12:59Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.