Understanding the Under-Coverage Bias in Uncertainty Estimation
- URL: http://arxiv.org/abs/2106.05515v1
- Date: Thu, 10 Jun 2021 06:11:55 GMT
- Title: Understanding the Under-Coverage Bias in Uncertainty Estimation
- Authors: Yu Bai, Song Mei, Huan Wang, Caiming Xiong
- Abstract summary: quantile regression tends to emphunder-cover than the desired coverage level in reality.
We prove that quantile regression suffers from an inherent under-coverage bias.
Our theory reveals that this under-coverage bias stems from a certain high-dimensional parameter estimation error.
- Score: 58.03725169462616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Estimating the data uncertainty in regression tasks is often done by learning
a quantile function or a prediction interval of the true label conditioned on
the input. It is frequently observed that quantile regression -- a vanilla
algorithm for learning quantiles with asymptotic guarantees -- tends to
\emph{under-cover} than the desired coverage level in reality. While various
fixes have been proposed, a more fundamental understanding of why this
under-coverage bias happens in the first place remains elusive.
In this paper, we present a rigorous theoretical study on the coverage of
uncertainty estimation algorithms in learning quantiles. We prove that quantile
regression suffers from an inherent under-coverage bias, in a vanilla setting
where we learn a realizable linear quantile function and there is more data
than parameters. More quantitatively, for $\alpha>0.5$ and small $d/n$, the
$\alpha$-quantile learned by quantile regression roughly achieves coverage
$\alpha - (\alpha-1/2)\cdot d/n$ regardless of the noise distribution, where
$d$ is the input dimension and $n$ is the number of training data. Our theory
reveals that this under-coverage bias stems from a certain high-dimensional
parameter estimation error that is not implied by existing theories on quantile
regression. Experiments on simulated and real data verify our theory and
further illustrate the effect of various factors such as sample size and model
capacity on the under-coverage bias in more practical setups.
Related papers
- Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Robust deep learning from weakly dependent data [0.0]
This paper considers robust deep learning from weakly dependent observations, with unbounded loss function and unbounded input/output.
We derive a relationship between these bounds and $r$, and when the data have moments of any order (that is $r=infty$), the convergence rate is close to some well-known results.
arXiv Detail & Related papers (2024-05-08T14:25:40Z) - Lazy Estimation of Variable Importance for Large Neural Networks [22.95405462638975]
We propose a fast and flexible method for approximating the reduced model with important inferential guarantees.
We demonstrate our method is fast and accurate under several data-generating regimes, and we demonstrate its real-world applicability on a seasonal climate forecasting example.
arXiv Detail & Related papers (2022-07-19T06:28:17Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z) - SLOE: A Faster Method for Statistical Inference in High-Dimensional
Logistic Regression [68.66245730450915]
We develop an improved method for debiasing predictions and estimating frequentist uncertainty for practical datasets.
Our main contribution is SLOE, an estimator of the signal strength with convergence guarantees that reduces the computation time of estimation and inference by orders of magnitude.
arXiv Detail & Related papers (2021-03-23T17:48:56Z) - Flexible Model Aggregation for Quantile Regression [92.63075261170302]
Quantile regression is a fundamental problem in statistical learning motivated by a need to quantify uncertainty in predictions.
We investigate methods for aggregating any number of conditional quantile models.
All of the models we consider in this paper can be fit using modern deep learning toolkits.
arXiv Detail & Related papers (2021-02-26T23:21:16Z) - Regularization Strategies for Quantile Regression [8.232258589877942]
We show that minimizing an expected pinball loss over a continuous distribution of quantiles is a good regularizer even when only predicting a specific quantile.
We show that lattice models enable regularizing the predicted distribution to a location-scale family.
arXiv Detail & Related papers (2021-02-09T21:10:35Z) - A Random Matrix Analysis of Random Fourier Features: Beyond the Gaussian
Kernel, a Precise Phase Transition, and the Corresponding Double Descent [85.77233010209368]
This article characterizes the exacts of random Fourier feature (RFF) regression, in the realistic setting where the number of data samples $n$ is all large and comparable.
This analysis also provides accurate estimates of training and test regression errors for large $n,p,N$.
arXiv Detail & Related papers (2020-06-09T02:05:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.