Learning Prediction Intervals for Regression: Generalization and
Calibration
- URL: http://arxiv.org/abs/2102.13625v1
- Date: Fri, 26 Feb 2021 17:55:30 GMT
- Title: Learning Prediction Intervals for Regression: Generalization and
Calibration
- Authors: Haoxian Chen, Ziyi Huang, Henry Lam, Huajie Qian, Haofeng Zhang
- Abstract summary: We study the generation of prediction intervals in regression for uncertainty quantification.
We use a general learning theory to characterize the optimality-feasibility tradeoff that encompasses Lipschitz continuity and VC-subgraph classes.
We empirically demonstrate the strengths of our interval generation and calibration algorithms in terms of testing performances compared to existing benchmarks.
- Score: 12.576284277353606
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the generation of prediction intervals in regression for uncertainty
quantification. This task can be formalized as an empirical constrained
optimization problem that minimizes the average interval width while
maintaining the coverage accuracy across data. We strengthen the existing
literature by studying two aspects of this empirical optimization. First is a
general learning theory to characterize the optimality-feasibility tradeoff
that encompasses Lipschitz continuity and VC-subgraph classes, which are
exemplified in regression trees and neural networks. Second is a calibration
machinery and the corresponding statistical theory to optimally select the
regularization parameter that manages this tradeoff, which bypasses the
overfitting issues in previous approaches in coverage attainment. We
empirically demonstrate the strengths of our interval generation and
calibration algorithms in terms of testing performances compared to existing
benchmarks.
Related papers
- Risk and cross validation in ridge regression with correlated samples [72.59731158970894]
We provide training examples for the in- and out-of-sample risks of ridge regression when the data points have arbitrary correlations.
We further extend our analysis to the case where the test point has non-trivial correlations with the training set, setting often encountered in time series forecasting.
We validate our theory across a variety of high dimensional data.
arXiv Detail & Related papers (2024-08-08T17:27:29Z) - PAC-Bayes Analysis for Recalibration in Classification [4.005483185111992]
We conduct a generalization analysis of the calibration error under the probably approximately correct (PAC) Bayes framework.
We then propose a generalization-aware recalibration algorithm based on our generalization theory.
arXiv Detail & Related papers (2024-06-10T12:53:13Z) - Conformalized Survival Distributions: A Generic Post-Process to Increase Calibration [6.868842871753991]
Discrimination and calibration represent two important properties of survival analysis.
With their distinct nature, it is hard for survival models to simultaneously optimize both of them.
This paper introduces a novel approach utilizing conformal regression that can improve a model's calibration without degrading discrimination.
arXiv Detail & Related papers (2024-05-12T20:27:34Z) - Multiclass Alignment of Confidence and Certainty for Network Calibration [10.15706847741555]
Recent studies reveal that deep neural networks (DNNs) are prone to making overconfident predictions.
We propose a new train-time calibration method, which features a simple, plug-and-play auxiliary loss known as multi-class alignment of predictive mean confidence and predictive certainty (MACC)
Our method achieves state-of-the-art calibration performance for both in-domain and out-domain predictions.
arXiv Detail & Related papers (2023-09-06T00:56:24Z) - A Large-Scale Study of Probabilistic Calibration in Neural Network
Regression [3.13468877208035]
We conduct the largest empirical study to date to assess the probabilistic calibration of neural networks.
We introduce novel differentiable recalibration and regularization methods, uncovering new insights into their effectiveness.
arXiv Detail & Related papers (2023-06-05T09:33:39Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Adaptive Dimension Reduction and Variational Inference for Transductive
Few-Shot Classification [2.922007656878633]
We propose a new clustering method based on Variational Bayesian inference, further improved by Adaptive Dimension Reduction.
Our proposed method significantly improves accuracy in the realistic unbalanced transductive setting on various Few-Shot benchmarks.
arXiv Detail & Related papers (2022-09-18T10:29:02Z) - Efficient and Differentiable Conformal Prediction with General Function
Classes [96.74055810115456]
We propose a generalization of conformal prediction to multiple learnable parameters.
We show that it achieves approximate valid population coverage and near-optimal efficiency within class.
Experiments show that our algorithm is able to learn valid prediction sets and improve the efficiency significantly.
arXiv Detail & Related papers (2022-02-22T18:37:23Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - Predicting Deep Neural Network Generalization with Perturbation Response
Curves [58.8755389068888]
We propose a new framework for evaluating the generalization capabilities of trained networks.
Specifically, we introduce two new measures for accurately predicting generalization gaps.
We attain better predictive scores than the current state-of-the-art measures on a majority of tasks in the Predicting Generalization in Deep Learning (PGDL) NeurIPS 2020 competition.
arXiv Detail & Related papers (2021-06-09T01:37:36Z) - Privacy Preserving Recalibration under Domain Shift [119.21243107946555]
We introduce a framework that abstracts out the properties of recalibration problems under differential privacy constraints.
We also design a novel recalibration algorithm, accuracy temperature scaling, that outperforms prior work on private datasets.
arXiv Detail & Related papers (2020-08-21T18:43:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.