Bias-Corrected Peaks-Over-Threshold Estimation of the CVaR
- URL: http://arxiv.org/abs/2103.05059v1
- Date: Mon, 8 Mar 2021 20:29:06 GMT
- Title: Bias-Corrected Peaks-Over-Threshold Estimation of the CVaR
- Authors: Dylan Troop, Fr\'ed\'eric Godin, Jia Yuan Yu
- Abstract summary: Conditional value-at-risk (CVaR) is a useful risk measure in fields such as machine learning, finance, insurance, energy, etc.
When measuring very extreme risk, the commonly used CVaR estimation method of sample averaging does not work well.
To mitigate this problem, the CVaR can be estimated by extrapolating above a lower threshold than the VaR.
- Score: 2.552459629685159
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The conditional value-at-risk (CVaR) is a useful risk measure in fields such
as machine learning, finance, insurance, energy, etc. When measuring very
extreme risk, the commonly used CVaR estimation method of sample averaging does
not work well due to limited data above the value-at-risk (VaR), the quantile
corresponding to the CVaR level. To mitigate this problem, the CVaR can be
estimated by extrapolating above a lower threshold than the VaR using a
generalized Pareto distribution (GPD), which is often referred to as the
peaks-over-threshold (POT) approach. This method often requires a very high
threshold to fit well, leading to high variance in estimation, and can induce
significant bias if the threshold is chosen too low. In this paper, we derive a
new expression for the GPD approximation error of the CVaR, a bias term induced
by the choice of threshold, as well as a bias correction method for the
estimated GPD parameters. This leads to the derivation of a new estimator for
the CVaR that we prove to be asymptotically unbiased. In a practical setting,
we show through experiments that our estimator provides a significant
performance improvement compared with competing CVaR estimators in finite
samples. As a consequence of our bias correction method, it is also shown that
a much lower threshold can be selected without introducing significant bias.
This allows a larger portion of data to be be used in CVaR estimation compared
with the typical POT approach, leading to more stable estimates. As secondary
results, a new estimator for a second-order parameter of heavy-tailed
distributions is derived, as well as a confidence interval for the CVaR which
enables quantifying the level of variability in our estimator.
Related papers
- Data-driven decision-making under uncertainty with entropic risk measure [5.407319151576265]
The entropic risk measure is widely used in high-stakes decision making to account for tail risks associated with an uncertain loss.
To debias the empirical entropic risk estimator, we propose a strongly consistent bootstrapping procedure.
We show that cross validation methods can result in significantly higher out-of-sample risk for the insurer if the bias in validation performance is not corrected for.
arXiv Detail & Related papers (2024-09-30T04:02:52Z) - Risk and cross validation in ridge regression with correlated samples [72.59731158970894]
We provide training examples for the in- and out-of-sample risks of ridge regression when the data points have arbitrary correlations.
We further extend our analysis to the case where the test point has non-trivial correlations with the training set, setting often encountered in time series forecasting.
We validate our theory across a variety of high dimensional data.
arXiv Detail & Related papers (2024-08-08T17:27:29Z) - Safe Deployment for Counterfactual Learning to Rank with Exposure-Based
Risk Minimization [63.93275508300137]
We introduce a novel risk-aware Counterfactual Learning To Rank method with theoretical guarantees for safe deployment.
Our experimental results demonstrate the efficacy of our proposed method, which is effective at avoiding initial periods of bad performance when little data is available.
arXiv Detail & Related papers (2023-04-26T15:54:23Z) - Off-Policy Risk Assessment in Markov Decision Processes [15.225153671736201]
We develop the first doubly robust (DR) estimator for the CDF of returns in Markov decision processes (MDPs)
This estimator enjoys significantly less variance and, when the model is well specified, achieves the Cramer-Rao variance lower bound.
We derive the first minimax lower bounds for off-policy CDF and risk estimation, which match our error bounds up to a constant factor.
arXiv Detail & Related papers (2022-09-21T15:40:59Z) - Risk-Averse No-Regret Learning in Online Convex Games [19.4481913405231]
We consider an online game with risk-averse agents whose goal is to learn optimal decisions that minimize the risk of incurring significantly high costs.
Since the distributions of the cost functions depend on the actions of all agents that are generally unobservable, the Conditional Value at Risk (CVaR) values of the costs are difficult to compute.
We propose a new online risk-averse learning algorithm that relies on one-point zeroth-order estimation of the CVaR gradients computed using CVaR values.
arXiv Detail & Related papers (2022-03-16T21:36:47Z) - Confidence intervals for the Cox model test error from cross-validation [91.3755431537592]
Cross-validation (CV) is one of the most widely used techniques in statistical learning for estimating the test error of a model.
Standard confidence intervals for test error using estimates from CV may have coverage below nominal levels.
One way to this issue is by estimating the mean squared error of the prediction error instead using nested CV.
arXiv Detail & Related papers (2022-01-26T06:40:43Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - Near-optimal inference in adaptive linear regression [60.08422051718195]
Even simple methods like least squares can exhibit non-normal behavior when data is collected in an adaptive manner.
We propose a family of online debiasing estimators to correct these distributional anomalies in at least squares estimation.
We demonstrate the usefulness of our theory via applications to multi-armed bandit, autoregressive time series estimation, and active learning with exploration.
arXiv Detail & Related papers (2021-07-05T21:05:11Z) - Enhanced Doubly Robust Learning for Debiasing Post-click Conversion Rate
Estimation [29.27760413892272]
Post-click conversion, as a strong signal indicating the user preference, is salutary for building recommender systems.
Currently, most existing methods utilize counterfactual learning to debias recommender systems.
We propose a novel double learning approach for the MRDR estimator, which can convert the error imputation into the general CVR estimation.
arXiv Detail & Related papers (2021-05-28T06:59:49Z) - Rao-Blackwellizing the Straight-Through Gumbel-Softmax Gradient
Estimator [93.05919133288161]
We show that the variance of the straight-through variant of the popular Gumbel-Softmax estimator can be reduced through Rao-Blackwellization.
This provably reduces the mean squared error.
We empirically demonstrate that this leads to variance reduction, faster convergence, and generally improved performance in two unsupervised latent variable models.
arXiv Detail & Related papers (2020-10-09T22:54:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.