Compromise-free Bayesian neural networks
- URL: http://arxiv.org/abs/2004.12211v3
- Date: Sat, 13 Jun 2020 12:03:28 GMT
- Title: Compromise-free Bayesian neural networks
- Authors: Kamran Javid, Will Handley, Mike Hobson, Anthony Lasenby
- Abstract summary: We numerically sample the full (non-Gaussian and multimodal) network posterior and obtain numerical estimates of the Bayesian evidence.
Networks with $ReLU$ activation functions have consistently higher evidences than those with $tanh$ functions.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We conduct a thorough analysis of the relationship between the out-of-sample
performance and the Bayesian evidence (marginal likelihood) of Bayesian neural
networks (BNNs), as well as looking at the performance of ensembles of BNNs,
both using the Boston housing dataset. Using the state-of-the-art in nested
sampling, we numerically sample the full (non-Gaussian and multimodal) network
posterior and obtain numerical estimates of the Bayesian evidence, considering
network models with up to 156 trainable parameters. The networks have between
zero and four hidden layers, either $\tanh$ or $ReLU$ activation functions, and
with and without hierarchical priors. The ensembles of BNNs are obtained by
determining the posterior distribution over networks, from the posterior
samples of individual BNNs re-weighted by the associated Bayesian evidence
values. There is good correlation between out-of-sample performance and
evidence, as well as a remarkable symmetry between the evidence versus model
size and out-of-sample performance versus model size planes. Networks with
$ReLU$ activation functions have consistently higher evidences than those with
$\tanh$ functions, and this is reflected in their out-of-sample performance.
Ensembling over architectures acts to further improve performance relative to
the individual BNNs.
Related papers
- Role of Locality and Weight Sharing in Image-Based Tasks: A Sample Complexity Separation between CNNs, LCNs, and FCNs [42.551773746803946]
Vision tasks are characterized by the properties of locality and translation invariance.
The superior performance of convolutional neural networks (CNNs) on these tasks is widely attributed to the inductive bias of locality and weight sharing baked into their architecture.
Existing attempts to quantify the statistical benefits of these biases in CNNs over locally connected neural networks (LCNs) and fully connected neural networks (FCNs) fall into one of the following categories.
arXiv Detail & Related papers (2024-03-23T03:57:28Z) - Individual Fairness in Bayesian Neural Networks [9.386341375741225]
We study Individual Fairness (IF) for Bayesian neural networks (BNNs)
We use bounds on statistical sampling over the input space and the relationship between adversarial and individual fairness to derive a framework for the robustness estimation of $epsilon$-$delta$-IF.
We find that BNNs trained by means of approximate Bayesian inference consistently tend to be markedly more individually fair than their deterministic counterparts.
arXiv Detail & Related papers (2023-04-21T09:12:14Z) - Constraining cosmological parameters from N-body simulations with
Variational Bayesian Neural Networks [0.0]
Multiplicative normalizing flows (MNFs) are a family of approximate posteriors for the parameters of BNNs.
We have compared MNFs with respect to the standard BNNs, and the flipout estimator.
MNFs provide more realistic predictive distribution closer to the true posterior mitigating the bias introduced by the variational approximation.
arXiv Detail & Related papers (2023-01-09T16:07:48Z) - Explicit Tradeoffs between Adversarial and Natural Distributional
Robustness [48.44639585732391]
In practice, models need to enjoy both types of robustness to ensure reliability.
In this work, we show that in fact, explicit tradeoffs exist between adversarial and natural distributional robustness.
arXiv Detail & Related papers (2022-09-15T19:58:01Z) - Batch-Ensemble Stochastic Neural Networks for Out-of-Distribution
Detection [55.028065567756066]
Out-of-distribution (OOD) detection has recently received much attention from the machine learning community due to its importance in deploying machine learning models in real-world applications.
In this paper we propose an uncertainty quantification approach by modelling the distribution of features.
We incorporate an efficient ensemble mechanism, namely batch-ensemble, to construct the batch-ensemble neural networks (BE-SNNs) and overcome the feature collapse problem.
We show that BE-SNNs yield superior performance on several OOD benchmarks, such as the Two-Moons dataset, the FashionMNIST vs MNIST dataset, FashionM
arXiv Detail & Related papers (2022-06-26T16:00:22Z) - Model Architecture Adaption for Bayesian Neural Networks [9.978961706999833]
We show a novel network architecture search (NAS) that optimize BNNs for both accuracy and uncertainty.
In our experiments, the searched models show comparable uncertainty ability and accuracy compared to the state-of-the-art (deep ensemble)
arXiv Detail & Related papers (2022-02-09T10:58:50Z) - Robustness of Bayesian Neural Networks to White-Box Adversarial Attacks [55.531896312724555]
Bayesian Networks (BNNs) are robust and adept at handling adversarial attacks by incorporating randomness.
We create our BNN model, called BNN-DenseNet, by fusing Bayesian inference (i.e., variational Bayes) to the DenseNet architecture.
An adversarially-trained BNN outperforms its non-Bayesian, adversarially-trained counterpart in most experiments.
arXiv Detail & Related papers (2021-11-16T16:14:44Z) - Adversarial Examples Detection with Bayesian Neural Network [57.185482121807716]
We propose a new framework to detect adversarial examples motivated by the observations that random components can improve the smoothness of predictors.
We propose a novel Bayesian adversarial example detector, short for BATer, to improve the performance of adversarial example detection.
arXiv Detail & Related papers (2021-05-18T15:51:24Z) - Siamese Neural Network with Joint Bayesian Model Structure for Speaker
Verification [54.96267179988487]
We propose a novel Siamese neural network (SiamNN) for speaker verification.
Joint distribution of samples is first formulated based on a joint Bayesian (JB) based generative model.
We further train the model parameters with the pair-wised samples as a binary discrimination task for speaker verification.
arXiv Detail & Related papers (2021-04-07T09:17:29Z) - Encoding the latent posterior of Bayesian Neural Networks for
uncertainty quantification [10.727102755903616]
We aim for efficient deep BNNs amenable to complex computer vision architectures.
We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer.
Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles.
arXiv Detail & Related papers (2020-12-04T19:50:09Z) - Exact posterior distributions of wide Bayesian neural networks [51.20413322972014]
We show that the exact BNN posterior converges (weakly) to the one induced by the GP limit of the prior.
For empirical validation, we show how to generate exact samples from a finite BNN on a small dataset via rejection sampling.
arXiv Detail & Related papers (2020-06-18T13:57:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.