Tailoring to the Tails: Risk Measures for Fine-Grained Tail Sensitivity
- URL: http://arxiv.org/abs/2208.03066v1
- Date: Fri, 5 Aug 2022 09:51:18 GMT
- Title: Tailoring to the Tails: Risk Measures for Fine-Grained Tail Sensitivity
- Authors: Christian Fr\"ohlich, Robert C. Williamson
- Abstract summary: Expected risk rearrangement (ERM) is at the core of machine learning systems.
We propose a general approach to construct risk measures which exhibit a desired tail sensitivity.
- Score: 10.482805367361818
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Expected risk minimization (ERM) is at the core of machine learning systems.
This means that the risk inherent in a loss distribution is summarized using a
single number - its average. In this paper, we propose a general approach to
construct risk measures which exhibit a desired tail sensitivity and may
replace the expectation operator in ERM. Our method relies on the specification
of a reference distribution with a desired tail behaviour, which is in a
one-to-one correspondence to a coherent upper probability. Any risk measure,
which is compatible with this upper probability, displays a tail sensitivity
which is finely tuned to the reference distribution. As a concrete example, we
focus on divergence risk measures based on f-divergence ambiguity sets, which
are a widespread tool used to foster distributional robustness of machine
learning systems. For instance, we show how ambiguity sets based on the
Kullback-Leibler divergence are intricately tied to the class of subexponential
random variables. We elaborate the connection of divergence risk measures and
rearrangement invariant Banach norms.
Related papers
- Data-driven decision-making under uncertainty with entropic risk measure [5.407319151576265]
The entropic risk measure is widely used in high-stakes decision making to account for tail risks associated with an uncertain loss.
To debias the empirical entropic risk estimator, we propose a strongly consistent bootstrapping procedure.
We show that cross validation methods can result in significantly higher out-of-sample risk for the insurer if the bias in validation performance is not corrected for.
arXiv Detail & Related papers (2024-09-30T04:02:52Z) - Data-Adaptive Tradeoffs among Multiple Risks in Distribution-Free Prediction [55.77015419028725]
We develop methods that permit valid control of risk when threshold and tradeoff parameters are chosen adaptively.
Our methodology supports monotone and nearly-monotone risks, but otherwise makes no distributional assumptions.
arXiv Detail & Related papers (2024-03-28T17:28:06Z) - Predictive Uncertainty Quantification via Risk Decompositions for Strictly Proper Scoring Rules [7.0549244915538765]
Uncertainty in predictive modeling often relies on ad hoc methods.
This paper introduces a theoretical approach to understanding uncertainty through statistical risks.
We show how to split pointwise risk into Bayes risk and excess risk.
arXiv Detail & Related papers (2024-02-16T14:40:22Z) - Likelihood Ratio Confidence Sets for Sequential Decision Making [51.66638486226482]
We revisit the likelihood-based inference principle and propose to use likelihood ratios to construct valid confidence sequences.
Our method is especially suitable for problems with well-specified likelihoods.
We show how to provably choose the best sequence of estimators and shed light on connections to online convex optimization.
arXiv Detail & Related papers (2023-11-08T00:10:21Z) - On the Variance, Admissibility, and Stability of Empirical Risk
Minimization [80.26309576810844]
Empirical Risk Minimization (ERM) with squared loss may attain minimax suboptimal error rates.
We show that under mild assumptions, the suboptimality of ERM must be due to large bias rather than variance.
We also show that our estimates imply stability of ERM, complementing the main result of Caponnetto and Rakhlin (2006) for non-Donsker classes.
arXiv Detail & Related papers (2023-05-29T15:25:48Z) - Policy Evaluation in Distributional LQR [70.63903506291383]
We provide a closed-form expression of the distribution of the random return.
We show that this distribution can be approximated by a finite number of random variables.
Using the approximate return distribution, we propose a zeroth-order policy gradient algorithm for risk-averse LQR.
arXiv Detail & Related papers (2023-03-23T20:27:40Z) - How does the Combined Risk Affect the Performance of Unsupervised Domain
Adaptation Approaches? [33.65954640678556]
Unsupervised domain adaptation (UDA) aims to train a target classifier with labeled samples from the source domain and unlabeled samples from the target domain.
E-MixNet employs enhanced mixup, a generic vicinal distribution, on the labeled source samples and pseudo-labeled target samples to calculate a proxy of the combined risk.
arXiv Detail & Related papers (2020-12-30T00:46:57Z) - Learning with risks based on M-location [6.903929927172917]
We study a new class of risks defined in terms of the location and deviation of the loss distribution.
The class is easily implemented as a wrapper around any smooth loss.
arXiv Detail & Related papers (2020-12-04T06:21:51Z) - Optimal Best-Arm Identification Methods for Tail-Risk Measures [9.128264779870538]
Conditional value-at-risk (CVaR) and value-at-risk (VaR) are popular tail-risk measures in finance and insurance industries.
We identify the smallest CVaR, VaR, or sum of CVaR and mean from amongst finitely that has smallest CVaR, VaR, or sum of CVaR and mean.
arXiv Detail & Related papers (2020-08-17T20:23:24Z) - Thompson Sampling Algorithms for Mean-Variance Bandits [97.43678751629189]
We develop Thompson Sampling-style algorithms for mean-variance MAB.
We also provide comprehensive regret analyses for Gaussian and Bernoulli bandits.
Our algorithms significantly outperform existing LCB-based algorithms for all risk tolerances.
arXiv Detail & Related papers (2020-02-01T15:33:50Z) - Orthogonal Statistical Learning [49.55515683387805]
We provide non-asymptotic excess risk guarantees for statistical learning in a setting where the population risk depends on an unknown nuisance parameter.
We show that if the population risk satisfies a condition called Neymanity, the impact of the nuisance estimation error on the excess risk bound achieved by the meta-algorithm is of second order.
arXiv Detail & Related papers (2019-01-25T02:21:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.