Evidential Turing Processes
- URL: http://arxiv.org/abs/2106.01216v1
- Date: Wed, 2 Jun 2021 15:09:20 GMT
- Title: Evidential Turing Processes
- Authors: Melih Kandemir, Abdullah Akg\"ul, Manuel Haussmann, Gozde Unal
- Abstract summary: We introduce an original combination of evidential deep learning, neural processes, and neural Turing machines.
We observe our method on three image classification benchmarks and two neural net architectures.
- Score: 11.021440340896786
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A probabilistic classifier with reliable predictive uncertainties i) fits
successfully to the target domain data, ii) provides calibrated class
probabilities in difficult regions of the target domain (e.g. class overlap),
and iii) accurately identifies queries coming out of the target domain and
reject them. We introduce an original combination of evidential deep learning,
neural processes, and neural Turing machines capable of providing all three
essential properties mentioned above for total uncertainty quantification. We
observe our method on three image classification benchmarks and two neural net
architectures to consistently give competitive or superior scores with respect
to multiple uncertainty quantification metrics against state-of-the-art methods
explicitly tailored to one or a few of them. Our unified solution delivers an
implementation-friendly and computationally efficient recipe for safety
clearance and provides intellectual economy to an investigation of algorithmic
roots of epistemic awareness in deep neural nets.
Related papers
- SMLE: Safe Machine Learning via Embedded Overapproximation [4.129133569151574]
We consider the task of training differentiable ML models guaranteed to satisfy designer-chosen properties.
This is very challenging, due to the computational complexity of rigorously verifying and enforcing compliance in modern neural models.
We provide an innovative approach based on three components: 1) a general, simple architecture enabling efficient verification with a conservative semantic.
We evaluate our approach on properties defined by linear inequalities in regression, and on mutually exclusive classes in multilabel classification.
arXiv Detail & Related papers (2024-09-30T17:19:57Z) - LoRA-Ensemble: Efficient Uncertainty Modelling for Self-attention Networks [52.46420522934253]
We introduce LoRA-Ensemble, a parameter-efficient deep ensemble method for self-attention networks.
By employing a single pre-trained self-attention network with weights shared across all members, we train member-specific low-rank matrices for the attention projections.
Our method exhibits superior calibration compared to explicit ensembles and achieves similar or better accuracy across various prediction tasks and datasets.
arXiv Detail & Related papers (2024-05-23T11:10:32Z) - Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - The Boundaries of Verifiable Accuracy, Robustness, and Generalisation in
Deep Learning [73.5095051707364]
We consider classical distribution-agnostic framework and algorithms minimising empirical risks.
We show that there is a large family of tasks for which computing and verifying ideal stable and accurate neural networks is extremely challenging.
arXiv Detail & Related papers (2023-09-13T16:33:27Z) - Semantic Strengthening of Neuro-Symbolic Learning [85.6195120593625]
Neuro-symbolic approaches typically resort to fuzzy approximations of a probabilistic objective.
We show how to compute this efficiently for tractable circuits.
We test our approach on three tasks: predicting a minimum-cost path in Warcraft, predicting a minimum-cost perfect matching, and solving Sudoku puzzles.
arXiv Detail & Related papers (2023-02-28T00:04:22Z) - Vertex-based reachability analysis for verifying ReLU deep neural
networks [3.5816079147181483]
We propose three novel reachability algorithms for verifying deep neural networks with ReLU activations.
Our experiments on the ACAS Xu problem show that the Exact Polytope Network Mapping (EPNM) reachability algorithm proposed in this work surpass the state-of-the-art results from the literature.
arXiv Detail & Related papers (2023-01-27T21:46:03Z) - A Bit More Bayesian: Domain-Invariant Learning with Uncertainty [111.22588110362705]
Domain generalization is challenging due to the domain shift and the uncertainty caused by the inaccessibility of target domain data.
In this paper, we address both challenges with a probabilistic framework based on variational Bayesian inference.
We derive domain-invariant representations and classifiers, which are jointly established in a two-layer Bayesian neural network.
arXiv Detail & Related papers (2021-05-09T21:33:27Z) - Multivariate Deep Evidential Regression [77.34726150561087]
A new approach with uncertainty-aware neural networks shows promise over traditional deterministic methods.
We discuss three issues with a proposed solution to extract aleatoric and epistemic uncertainties from regression-based neural networks.
arXiv Detail & Related papers (2021-04-13T12:20:18Z) - Data-Driven Assessment of Deep Neural Networks with Random Input
Uncertainty [14.191310794366075]
We develop a data-driven optimization-based method capable of simultaneously certifying the safety of network outputs and localizing them.
We experimentally demonstrate the efficacy and tractability of the method on a deep ReLU network.
arXiv Detail & Related papers (2020-10-02T19:13:35Z) - Reachable Sets of Classifiers and Regression Models: (Non-)Robustness
Analysis and Robust Training [1.0878040851638]
We analyze and enhance robustness properties of both classifiers and regression models.
Specifically, we verify (non-)robustness, propose a robust training procedure, and show that our approach outperforms adversarial attacks.
Second, we provide techniques to distinguish between reliable and non-reliable predictions for unlabeled inputs, to quantify the influence of each feature on a prediction, and compute a feature ranking.
arXiv Detail & Related papers (2020-07-28T10:58:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.