Are Bayesian neural networks intrinsically good at out-of-distribution
detection?
- URL: http://arxiv.org/abs/2107.12248v1
- Date: Mon, 26 Jul 2021 14:53:14 GMT
- Title: Are Bayesian neural networks intrinsically good at out-of-distribution
detection?
- Authors: Christian Henning, Francesco D'Angelo, Benjamin F. Grewe
- Abstract summary: It is widely assumed that Bayesian neural networks (BNN) are well suited for out-of-distribution (OOD) detection.
In this paper, we provide empirical evidence that proper Bayesian inference with common neural network architectures does not necessarily lead to good OOD detection.
- Score: 4.297070083645049
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The need to avoid confident predictions on unfamiliar data has sparked
interest in out-of-distribution (OOD) detection. It is widely assumed that
Bayesian neural networks (BNN) are well suited for this task, as the endowed
epistemic uncertainty should lead to disagreement in predictions on outliers.
In this paper, we question this assumption and provide empirical evidence that
proper Bayesian inference with common neural network architectures does not
necessarily lead to good OOD detection. To circumvent the use of approximate
inference, we start by studying the infinite-width case, where Bayesian
inference can be exact considering the corresponding Gaussian process.
Strikingly, the kernels induced under common architectural choices lead to
uncertainties that do not reflect the underlying data generating process and
are therefore unsuited for OOD detection. Finally, we study finite-width
networks using HMC, and observe OOD behavior that is consistent with the
infinite-width case. Overall, our study discloses fundamental problems when
naively using BNNs for OOD detection and opens interesting avenues for future
research.
Related papers
- Be Bayesian by Attachments to Catch More Uncertainty [27.047781689062944]
We propose a new Bayesian Neural Network with an Attached structure (ABNN) to catch more uncertainty from out-of-distribution (OOD) data.
ABNN is composed of an expectation module and several distribution modules.
arXiv Detail & Related papers (2023-10-19T07:28:39Z) - Deep Neural Networks Tend To Extrapolate Predictably [51.303814412294514]
neural network predictions tend to be unpredictable and overconfident when faced with out-of-distribution (OOD) inputs.
We observe that neural network predictions often tend towards a constant value as input data becomes increasingly OOD.
We show how one can leverage our insights in practice to enable risk-sensitive decision-making in the presence of OOD inputs.
arXiv Detail & Related papers (2023-10-02T03:25:32Z) - Energy-based Out-of-Distribution Detection for Graph Neural Networks [76.0242218180483]
We propose a simple, powerful and efficient OOD detection model for GNN-based learning on graphs, which we call GNNSafe.
GNNSafe achieves up to $17.0%$ AUROC improvement over state-of-the-arts and it could serve as simple yet strong baselines in such an under-developed area.
arXiv Detail & Related papers (2023-02-06T16:38:43Z) - Out of Distribution Data Detection Using Dropout Bayesian Neural
Networks [29.84998820573774]
We first show how previous attempts to leverage the randomized embeddings induced by the intermediate layers of a dropout BNN can fail due to the distance metric used.
We introduce an alternative approach to measuring embedding uncertainty, justify its use theoretically, and demonstrate how incorporating embedding uncertainty improves OOD data identification across three tasks: image classification, language classification, and malware detection.
arXiv Detail & Related papers (2022-02-18T02:23:43Z) - Uncertainty-based out-of-distribution detection requires suitable
function space priors [1.90365714903665]
We show that proper Bayesian inference with function space priors induced by neural networks does not necessarily lead to good OOD detection.
Desirable function space properties can be encoded in the prior in weight space, however, this currently only applies to a specified subset of the domain.
arXiv Detail & Related papers (2021-10-12T14:11:37Z) - Provably Robust Detection of Out-of-distribution Data (almost) for free [124.14121487542613]
Deep neural networks are known to produce highly overconfident predictions on out-of-distribution (OOD) data.
In this paper we propose a novel method where from first principles we combine a certifiable OOD detector with a standard classifier into an OOD aware classifier.
In this way we achieve the best of two worlds: certifiably adversarially robust OOD detection, even for OOD samples close to the in-distribution, without loss in prediction accuracy and close to state-of-the-art OOD detection performance for non-manipulated OOD data.
arXiv Detail & Related papers (2021-06-08T11:40:49Z) - Statistical Testing for Efficient Out of Distribution Detection in Deep
Neural Networks [26.0303701309125]
This paper frames the Out Of Distribution (OOD) detection problem in Deep Neural Networks as a statistical hypothesis testing problem.
We build on this framework to suggest a novel OOD procedure based on low-order statistics.
Our method achieves comparable or better than state-of-the-art results on well-accepted OOD benchmarks without retraining the network parameters.
arXiv Detail & Related papers (2021-02-25T16:14:47Z) - Learn what you can't learn: Regularized Ensembles for Transductive
Out-of-distribution Detection [76.39067237772286]
We show that current out-of-distribution (OOD) detection algorithms for neural networks produce unsatisfactory results in a variety of OOD detection scenarios.
This paper studies how such "hard" OOD scenarios can benefit from adjusting the detection method after observing a batch of the test data.
We propose a novel method that uses an artificial labeling scheme for the test data and regularization to obtain ensembles of models that produce contradictory predictions only on the OOD samples in a test batch.
arXiv Detail & Related papers (2020-12-10T16:55:13Z) - The Hidden Uncertainty in a Neural Networks Activations [105.4223982696279]
The distribution of a neural network's latent representations has been successfully used to detect out-of-distribution (OOD) data.
This work investigates whether this distribution correlates with a model's epistemic uncertainty, thus indicating its ability to generalise to novel inputs.
arXiv Detail & Related papers (2020-12-05T17:30:35Z) - A Survey on Assessing the Generalization Envelope of Deep Neural
Networks: Predictive Uncertainty, Out-of-distribution and Adversarial Samples [77.99182201815763]
Deep Neural Networks (DNNs) achieve state-of-the-art performance on numerous applications.
It is difficult to tell beforehand if a DNN receiving an input will deliver the correct output since their decision criteria are usually nontransparent.
This survey connects the three fields within the larger framework of investigating the generalization performance of machine learning methods and in particular DNNs.
arXiv Detail & Related papers (2020-08-21T09:12:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.