Correcting Classification: A Bayesian Framework Using Explanation
Feedback to Improve Classification Abilities
- URL: http://arxiv.org/abs/2105.02653v1
- Date: Thu, 29 Apr 2021 13:59:21 GMT
- Title: Correcting Classification: A Bayesian Framework Using Explanation
Feedback to Improve Classification Abilities
- Authors: Yanzhe Bekkemoen, Helge Langseth
- Abstract summary: Explanations are social, meaning they are a transfer of knowledge through interactions.
We overcome these difficulties by training a Bayesian convolutional neural network (CNN) that uses explanation feedback.
Our proposed method utilizes this feedback for fine-tuning to correct the model such that the explanations and classifications improve.
- Score: 2.0931163605360115
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural networks (NNs) have shown high predictive performance, however, with
shortcomings. Firstly, the reasons behind the classifications are not fully
understood. Several explanation methods have been developed, but they do not
provide mechanisms for users to interact with the explanations. Explanations
are social, meaning they are a transfer of knowledge through interactions.
Nonetheless, current explanation methods contribute only to one-way
communication. Secondly, NNs tend to be overconfident, providing unreasonable
uncertainty estimates on out-of-distribution observations. We overcome these
difficulties by training a Bayesian convolutional neural network (CNN) that
uses explanation feedback. After training, the model presents explanations of
training sample classifications to an annotator. Based on the provided
information, the annotator can accept or reject the explanations by providing
feedback. Our proposed method utilizes this feedback for fine-tuning to correct
the model such that the explanations and classifications improve. We use
existing CNN architectures to demonstrate the method's effectiveness on one toy
dataset (decoy MNIST) and two real-world datasets (Dogs vs. Cats and ISIC skin
cancer). The experiments indicate that few annotated explanations and
fine-tuning epochs are needed to improve the model and predictive performance,
making the model more trustworthy and understandable.
Related papers
- Sparsifying Bayesian neural networks with latent binary variables and
normalizing flows [10.865434331546126]
We will consider two extensions to the latent binary Bayesian neural networks (LBBNN) method.
Firstly, by using the local reparametrization trick (LRT) to sample the hidden units directly, we get a more computationally efficient algorithm.
More importantly, by using normalizing flows on the variational posterior distribution of the LBBNN parameters, the network learns a more flexible variational posterior distribution than the mean field Gaussian.
arXiv Detail & Related papers (2023-05-05T09:40:28Z) - Benign Overfitting for Two-layer ReLU Convolutional Neural Networks [60.19739010031304]
We establish algorithm-dependent risk bounds for learning two-layer ReLU convolutional neural networks with label-flipping noise.
We show that, under mild conditions, the neural network trained by gradient descent can achieve near-zero training loss and Bayes optimal test risk.
arXiv Detail & Related papers (2023-03-07T18:59:38Z) - Look beyond labels: Incorporating functional summary information in
Bayesian neural networks [11.874130244353253]
We present a simple approach to incorporate summary information about the predicted probability.
The available summary information is incorporated as augmented data and modeled with a Dirichlet process.
We show how the method can inform the model about task difficulty or class imbalance.
arXiv Detail & Related papers (2022-07-04T07:06:45Z) - Augmenting Neural Networks with Priors on Function Values [22.776982718042962]
Prior knowledge of function values is often available in the natural sciences.
BNNs enable the user to specify prior information only on the neural network weights, not directly on the function values.
We develop an approach to augment BNNs with prior information on the function values themselves.
arXiv Detail & Related papers (2022-02-10T02:24:15Z) - Training on Test Data with Bayesian Adaptation for Covariate Shift [96.3250517412545]
Deep neural networks often make inaccurate predictions with unreliable uncertainty estimates.
We derive a Bayesian model that provides for a well-defined relationship between unlabeled inputs under distributional shift and model parameters.
We show that our method improves both accuracy and uncertainty estimation.
arXiv Detail & Related papers (2021-09-27T01:09:08Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Learning Invariances in Neural Networks [51.20867785006147]
We show how to parameterize a distribution over augmentations and optimize the training loss simultaneously with respect to the network parameters and augmentation parameters.
We can recover the correct set and extent of invariances on image classification, regression, segmentation, and molecular property prediction from a large space of augmentations.
arXiv Detail & Related papers (2020-10-22T17:18:48Z) - Why have a Unified Predictive Uncertainty? Disentangling it using Deep
Split Ensembles [39.29536042476913]
Understanding and quantifying uncertainty in black box Neural Networks (NNs) is critical when deployed in real-world settings such as healthcare.
We propose a conceptually simple non-Bayesian approach, deep split ensemble, to disentangle the predictive uncertainties.
arXiv Detail & Related papers (2020-09-25T19:15:26Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.