Conformal Prediction Intervals for Neural Networks Using Cross
Validation
- URL: http://arxiv.org/abs/2006.16941v1
- Date: Tue, 30 Jun 2020 16:23:28 GMT
- Title: Conformal Prediction Intervals for Neural Networks Using Cross
Validation
- Authors: Saeed Khaki and Dan Nettleton
- Abstract summary: Neural networks are among the most powerful nonlinear models used to address supervised learning problems.
We propose the $k$-fold prediction interval method to construct prediction intervals for neural networks based on $k$-fold cross validation.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural networks are among the most powerful nonlinear models used to address
supervised learning problems. Similar to most machine learning algorithms,
neural networks produce point predictions and do not provide any prediction
interval which includes an unobserved response value with a specified
probability. In this paper, we proposed the $k$-fold prediction interval method
to construct prediction intervals for neural networks based on $k$-fold cross
validation. Simulation studies and analysis of 10 real datasets are used to
compare the finite-sample properties of the prediction intervals produced by
the proposed method and the split conformal (SC) method. The results suggest
that the proposed method tends to produce narrower prediction intervals
compared to the SC method while maintaining the same coverage probability. Our
experimental results also reveal that the proposed $k$-fold prediction interval
method produces effective prediction intervals and is especially advantageous
relative to competing approaches when the number of training observations is
limited.
Related papers
- Uncertainty Quantification via Stable Distribution Propagation [60.065272548502]
We propose a new approach for propagating stable probability distributions through neural networks.
Our method is based on local linearization, which we show to be an optimal approximation in terms of total variation distance for the ReLU non-linearity.
arXiv Detail & Related papers (2024-02-13T09:40:19Z) - Reliable Prediction Intervals with Regression Neural Networks [1.569545894307769]
We propose an extension to conventional regression Neural Networks (NNs) for replacing the point predictions they produce with prediction intervals that satisfy a required level of confidence.
Our approach follows a novel machine learning framework, called Conformal Prediction (CP), for assigning reliable confidence measures to predictions.
We evaluate the proposed method on four benchmark datasets and on the problem of predicting Total Electron Content (TEC), which is an important parameter in trans-ionospheric links.
arXiv Detail & Related papers (2023-12-15T08:39:02Z) - Prediction-Powered Inference [68.97619568620709]
Prediction-powered inference is a framework for performing valid statistical inference when an experimental dataset is supplemented with predictions from a machine-learning system.
The framework yields simple algorithms for computing provably valid confidence intervals for quantities such as means, quantiles, and linear and logistic regression coefficients.
Prediction-powered inference could enable researchers to draw valid and more data-efficient conclusions using machine learning.
arXiv Detail & Related papers (2023-01-23T18:59:28Z) - Scalable computation of prediction intervals for neural networks via
matrix sketching [79.44177623781043]
Existing algorithms for uncertainty estimation require modifying the model architecture and training procedure.
This work proposes a new algorithm that can be applied to a given trained neural network and produces approximate prediction intervals.
arXiv Detail & Related papers (2022-05-06T13:18:31Z) - Probabilistic AutoRegressive Neural Networks for Accurate Long-range
Forecasting [6.295157260756792]
We introduce the Probabilistic AutoRegressive Neural Networks (PARNN)
PARNN is capable of handling complex time series data exhibiting non-stationarity, nonlinearity, non-seasonality, long-range dependence, and chaotic patterns.
We evaluate the performance of PARNN against standard statistical, machine learning, and deep learning models, including Transformers, NBeats, and DeepAR.
arXiv Detail & Related papers (2022-04-01T17:57:36Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Probabilistic Forecasting with Generative Networks via Scoring Rule
Minimization [5.5643498845134545]
We use generative neural networks to parametrize distributions on high-dimensional spaces by transforming draws from a latent variable.
We train generative networks to minimize a predictive-sequential (or prequential) scoring rule on a recorded temporal sequence of the phenomenon of interest.
Our method outperforms state-of-the-art adversarial approaches, especially in probabilistic calibration.
arXiv Detail & Related papers (2021-12-15T15:51:12Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Quantifying Uncertainty in Deep Spatiotemporal Forecasting [67.77102283276409]
We describe two types of forecasting problems: regular grid-based and graph-based.
We analyze UQ methods from both the Bayesian and the frequentist point view, casting in a unified framework via statistical decision theory.
Through extensive experiments on real-world road network traffic, epidemics, and air quality forecasting tasks, we reveal the statistical computational trade-offs for different UQ methods.
arXiv Detail & Related papers (2021-05-25T14:35:46Z) - Interpretable Machines: Constructing Valid Prediction Intervals with
Random Forests [0.0]
An important issue when using Machine Learning algorithms in recent research is the lack of interpretability.
A contribution to this gap for the Random Forest Regression Learner is presented here.
Several parametric and non-parametric prediction intervals are provided for Random Forest point predictions.
A thorough investigation through Monte-Carlo simulation is conducted evaluating the performance of the proposed methods.
arXiv Detail & Related papers (2021-03-09T23:05:55Z) - Prediction intervals for Deep Neural Networks [0.0]
We adapt the randomized trees method originally developed for random forests to construct ensembles of neural networks.
The extra-randomness introduced in the ensemble reduces the variance of the predictions and yields gains in out-of-sample accuracy.
arXiv Detail & Related papers (2020-10-08T15:11:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.