Quantifying Uncertainty in Deep Learning Approaches to Radio Galaxy
Classification
- URL: http://arxiv.org/abs/2201.01203v1
- Date: Tue, 4 Jan 2022 15:31:18 GMT
- Title: Quantifying Uncertainty in Deep Learning Approaches to Radio Galaxy
Classification
- Authors: Devina Mohan, Anna M. M. Scaife, Fiona Porter, Mike Walmsley, Micah
Bowles
- Abstract summary: We show that the level of model posterior variance for individual test samples is correlated with human uncertainty when labelling radio galaxies.
We explore the model performance and uncertainty calibration for a variety of different weight priors and suggest that a sparse prior produces more well-calibrated uncertainty estimates.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work we use variational inference to quantify the degree of
uncertainty in deep learning model predictions of radio galaxy classification.
We show that the level of model posterior variance for individual test samples
is correlated with human uncertainty when labelling radio galaxies. We explore
the model performance and uncertainty calibration for a variety of different
weight priors and suggest that a sparse prior produces more well-calibrated
uncertainty estimates. Using the posterior distributions for individual
weights, we show that we can prune 30% of the fully-connected layer weights
without significant loss of performance by removing the weights with the lowest
signal-to-noise ratio (SNR). We demonstrate that a larger degree of pruning can
be achieved using a Fisher information based ranking, but we note that both
pruning methods affect the uncertainty calibration for Fanaroff-Riley type I
and type II radio galaxies differently. Finally we show that, like other work
in this field, we experience a cold posterior effect, whereby the posterior
must be down-weighted to achieve good predictive performance. We examine
whether adapting the cost function to accommodate model misspecification can
compensate for this effect, but find that it does not make a significant
difference. We also examine the effect of principled data augmentation and find
that this improves upon the baseline but also does not compensate for the
observed effect. We interpret this as the cold posterior effect being due to
the overly effective curation of our training sample leading to likelihood
misspecification, and raise this as a potential issue for Bayesian deep
learning approaches to radio galaxy classification in future.
Related papers
- Differentiable Pareto-Smoothed Weighting for High-Dimensional Heterogeneous Treatment Effect Estimation [0.6906005491572401]
We develop a numerically robust estimator by weighted representation learning.
Our experimental results show that by effectively correcting the weight values, our proposed method outperforms the existing ones.
arXiv Detail & Related papers (2024-04-26T15:34:04Z) - Posterior samples of source galaxies in strong gravitational lenses with
score-based priors [107.52670032376555]
We use a score-based model to encode the prior for the inference of undistorted images of background galaxies.
We show how the balance between the likelihood and the prior meet our expectations in an experiment with out-of-distribution data.
arXiv Detail & Related papers (2022-11-07T19:00:42Z) - Propagating Variational Model Uncertainty for Bioacoustic Call Label
Smoothing [15.929064190849665]
We focus on using the predictive uncertainty signal calculated by Bayesian neural networks to guide learning in the self-same task the model is being trained on.
Not opting for costly Monte Carlo sampling of weights, we propagate the approximate hidden variance in an end-to-end manner.
We show that, through the explicit usage of the uncertainty in the loss calculation, the variational model is led to improved predictive and calibration performance.
arXiv Detail & Related papers (2022-10-19T13:04:26Z) - Monotonicity and Double Descent in Uncertainty Estimation with Gaussian
Processes [52.92110730286403]
It is commonly believed that the marginal likelihood should be reminiscent of cross-validation metrics and that both should deteriorate with larger input dimensions.
We prove that by tuning hyper parameters, the performance, as measured by the marginal likelihood, improves monotonically with the input dimension.
We also prove that cross-validation metrics exhibit qualitatively different behavior that is characteristic of double descent.
arXiv Detail & Related papers (2022-10-14T08:09:33Z) - Neural Importance Sampling for Rapid and Reliable Gravitational-Wave
Inference [59.040209568168436]
We first generate a rapid proposal for the Bayesian posterior using neural networks, and then attach importance weights based on the underlying likelihood and prior.
This provides (1) a corrected posterior free from network inaccuracies, (2) a performance diagnostic (the sample efficiency) for assessing the proposal and identifying failure cases, and (3) an unbiased estimate of the Bayesian evidence.
We carry out a large study analyzing 42 binary black hole mergers observed by LIGO and Virgo with the SEOBNRv4PHM and IMRPhenomHMXP waveform models.
arXiv Detail & Related papers (2022-10-11T18:00:02Z) - Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification [74.62203971625173]
Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
arXiv Detail & Related papers (2022-08-05T01:23:54Z) - On Calibrated Model Uncertainty in Deep Learning [0.0]
We extend the approximate inference for the loss-calibrated Bayesian framework to dropweights based Bayesian neural networks.
We show that decisions informed by loss-calibrated uncertainty can improve diagnostic performance to a greater extent than straightforward alternatives.
arXiv Detail & Related papers (2022-06-15T20:16:32Z) - Benign-Overfitting in Conditional Average Treatment Effect Prediction
with Linear Regression [14.493176427999028]
We study the benign overfitting theory in the prediction of the conditional average treatment effect (CATE) with linear regression models.
We show that the T-learner fails to achieve the consistency except the random assignment, while the IPW-learner converges the risk to zero if the propensity score is known.
arXiv Detail & Related papers (2022-02-10T18:51:52Z) - Weight Pruning and Uncertainty in Radio Galaxy Classification [0.0]
We use variational inference to quantify the degree of uncertainty in model predictions of radio galaxy classification.
We show that the level of model posterior variance for individual test samples is correlated with human uncertainty when labelling radio galaxies.
arXiv Detail & Related papers (2021-11-23T05:01:27Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.