Randomized ReLU Activation for Uncertainty Estimation of Deep Neural
Networks
- URL: http://arxiv.org/abs/2107.07197v1
- Date: Thu, 15 Jul 2021 08:54:41 GMT
- Title: Randomized ReLU Activation for Uncertainty Estimation of Deep Neural
Networks
- Authors: Yufeng Xia, Jun Zhang, Zhiqiang Gong, Tingsong Jiang and Wen Yao
- Abstract summary: Deep neural networks (DNNs) have successfully learned useful data representations in various tasks.
Deep Ensemble is widely considered the state-of-the-art method for uncertainty estimation, but it is very expensive to train and test.
We introduce Randomized ReLU Activation framework to get more diverse predictions in less time.
- Score: 8.541875999755593
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Deep neural networks (DNNs) have successfully learned useful data
representations in various tasks, however, assessing the reliability of these
representations remains a challenge. Deep Ensemble is widely considered the
state-of-the-art method for uncertainty estimation, but it is very expensive to
train and test. MC-Dropout is another alternative method, which is less
expensive but lacks the diversity of predictions. To get more diverse
predictions in less time, we introduce Randomized ReLU Activation (RRA)
framework. Under the framework, we propose two strategies, MC-DropReLU and
MC-RReLU, to estimate uncertainty. Instead of randomly dropping some neurons of
the network as in MC-Dropout, the RRA framework adds randomness to the
activation function module, making the outputs diverse. As far as we know, this
is the first attempt to add randomness to the activation function module to
generate predictive uncertainty. We analyze and compare the output diversity of
MC-Dropout and our method from the variance perspective and obtain the
relationship between the hyperparameters and output diversity in the two
methods. Moreover, our method is simple to implement and does not need to
modify the existing model. We experimentally validate the RRA framework on
three widely used datasets, CIFAR10, CIFAR100, and TinyImageNet. The
experiments demonstrate that our method has competitive performance but is more
favorable in training time and memory requirements.
Related papers
- OPONeRF: One-Point-One NeRF for Robust Neural Rendering [70.56874833759241]
We propose a One-Point-One NeRF (OPONeRF) framework for robust scene rendering.
Small but unpredictable perturbations such as object movements, light changes and data contaminations broadly exist in real-life 3D scenes.
Experimental results show that our OPONeRF outperforms state-of-the-art NeRFs on various evaluation metrics.
arXiv Detail & Related papers (2024-09-30T07:49:30Z) - Uncertainty-aware Multi-modal Learning via Cross-modal Random Network
Prediction [22.786774541083652]
We propose a new Uncertainty-aware Multi-modal Learner that estimates uncertainty by measuring feature density via Cross-modal Random Network Prediction (CRNP)
CRNP is designed to require little adaptation to translate between different prediction tasks, while having a stable training process.
arXiv Detail & Related papers (2022-07-22T03:00:10Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Trustworthy Multimodal Regression with Mixture of Normal-inverse Gamma
Distributions [91.63716984911278]
We introduce a novel Mixture of Normal-Inverse Gamma distributions (MoNIG) algorithm, which efficiently estimates uncertainty in principle for adaptive integration of different modalities and produces a trustworthy regression result.
Experimental results on both synthetic and different real-world data demonstrate the effectiveness and trustworthiness of our method on various multimodal regression tasks.
arXiv Detail & Related papers (2021-11-11T14:28:12Z) - Contextual Dropout: An Efficient Sample-Dependent Dropout Module [60.63525456640462]
Dropout has been demonstrated as a simple and effective module to regularize the training process of deep neural networks.
We propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module.
Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation.
arXiv Detail & Related papers (2021-03-06T19:30:32Z) - A Novel Regression Loss for Non-Parametric Uncertainty Optimization [7.766663822644739]
Quantification of uncertainty is one of the most promising approaches to establish safe machine learning.
One of the most commonly used approaches so far is Monte Carlo dropout, which is computationally cheap and easy to apply in practice.
We propose a new objective, referred to as second-moment loss ( UCI), to address this issue.
arXiv Detail & Related papers (2021-01-07T19:12:06Z) - Multi-Loss Sub-Ensembles for Accurate Classification with Uncertainty
Estimation [1.2891210250935146]
We propose an efficient method for uncertainty estimation in deep neural networks (DNNs) achieving high accuracy.
We keep our inference time relatively low by leveraging the advantage proposed by the Deep-Sub-Ensembles method.
Our results show improved accuracy on the classification task and competitive results on several uncertainty measures.
arXiv Detail & Related papers (2020-10-05T10:59:11Z) - Beyond Point Estimate: Inferring Ensemble Prediction Variation from
Neuron Activation Strength in Recommender Systems [21.392694985689083]
Ensemble method is one state-of-the-art benchmark for prediction uncertainty estimation.
We observe that prediction variations come from various randomness sources.
We propose to infer prediction variation from neuron activation strength and demonstrate the strong prediction power from activation strength features.
arXiv Detail & Related papers (2020-08-17T00:08:27Z) - RAIN: A Simple Approach for Robust and Accurate Image Classification
Networks [156.09526491791772]
It has been shown that the majority of existing adversarial defense methods achieve robustness at the cost of sacrificing prediction accuracy.
This paper proposes a novel preprocessing framework, which we term Robust and Accurate Image classificatioN(RAIN)
RAIN applies randomization over inputs to break the ties between the model forward prediction path and the backward gradient path, thus improving the model robustness.
We conduct extensive experiments on the STL10 and ImageNet datasets to verify the effectiveness of RAIN against various types of adversarial attacks.
arXiv Detail & Related papers (2020-04-24T02:03:56Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.