Holdouts set for predictive model updating
- URL: http://arxiv.org/abs/2202.06374v4
- Date: Mon, 31 Jul 2023 11:39:21 GMT
- Title: Holdouts set for predictive model updating
- Authors: Sami Haidar-Wehbe, Samuel R Emerson, Louis J M Aslett, James Liley
- Abstract summary: Updating risk scores can lead to biased risk estimates.
We propose using a holdout set' - a subset of the population that does not receive interventions guided by the risk score.
We prove that this approach enables total costs to grow at a rate $Oleft(N2/3right)$ for a population of size $N$, and argue that in general circumstances there is no competitive alternative.
- Score: 0.9749560288448114
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In complex settings, such as healthcare, predictive risk scores play an
increasingly crucial role in guiding interventions. However, directly updating
risk scores used to guide intervention can lead to biased risk estimates. To
address this, we propose updating using a `holdout set' - a subset of the
population that does not receive interventions guided by the risk score.
Striking a balance in the size of the holdout set is essential, to ensure good
performance of the updated risk score whilst minimising the number of held out
samples. We prove that this approach enables total costs to grow at a rate
$O\left(N^{2/3}\right)$ for a population of size $N$, and argue that in general
circumstances there is no competitive alternative. By defining an appropriate
loss function, we describe conditions under which an optimal holdout size (OHS)
can be readily identified, and introduce parametric and semi-parametric
algorithms for OHS estimation, demonstrating their use on a recent risk score
for pre-eclampsia. Based on these results, we make the case that a holdout set
is a safe, viable and easily implemented means to safely update predictive risk
scores.
Related papers
- Data-driven decision-making under uncertainty with entropic risk measure [5.407319151576265]
The entropic risk measure is widely used in high-stakes decision making to account for tail risks associated with an uncertain loss.
To debias the empirical entropic risk estimator, we propose a strongly consistent bootstrapping procedure.
We show that cross validation methods can result in significantly higher out-of-sample risk for the insurer if the bias in validation performance is not corrected for.
arXiv Detail & Related papers (2024-09-30T04:02:52Z) - Model-Based Epistemic Variance of Values for Risk-Aware Policy Optimization [59.758009422067]
We consider the problem of quantifying uncertainty over expected cumulative rewards in model-based reinforcement learning.
We propose a new uncertainty Bellman equation (UBE) whose solution converges to the true posterior variance over values.
We introduce a general-purpose policy optimization algorithm, Q-Uncertainty Soft Actor-Critic (QU-SAC) that can be applied for either risk-seeking or risk-averse policy optimization.
arXiv Detail & Related papers (2023-12-07T15:55:58Z) - Safe Deployment for Counterfactual Learning to Rank with Exposure-Based
Risk Minimization [63.93275508300137]
We introduce a novel risk-aware Counterfactual Learning To Rank method with theoretical guarantees for safe deployment.
Our experimental results demonstrate the efficacy of our proposed method, which is effective at avoiding initial periods of bad performance when little data is available.
arXiv Detail & Related papers (2023-04-26T15:54:23Z) - RAP: Risk-Aware Prediction for Robust Planning [21.83865866611308]
We introduce a new prediction objective to learn a risk-biased distribution over trajectories.
This reduces the sample complexity of the risk estimation during online planning.
arXiv Detail & Related papers (2022-10-04T04:19:15Z) - Off-Policy Risk Assessment in Markov Decision Processes [15.225153671736201]
We develop the first doubly robust (DR) estimator for the CDF of returns in Markov decision processes (MDPs)
This estimator enjoys significantly less variance and, when the model is well specified, achieves the Cramer-Rao variance lower bound.
We derive the first minimax lower bounds for off-policy CDF and risk estimation, which match our error bounds up to a constant factor.
arXiv Detail & Related papers (2022-09-21T15:40:59Z) - A Risk-Sensitive Approach to Policy Optimization [21.684251937825234]
Standard deep reinforcement learning (DRL) aims to maximize expected reward, considering collected experiences equally in formulating a policy.
We propose a more direct approach whereby risk-sensitive objectives, specified in terms of the cumulative distribution function (CDF) of the distribution of full-episode rewards, are optimized.
We demonstrate that the use of moderately "pessimistic" risk profiles, which emphasize scenarios where the agent performs poorly, leads to enhanced exploration and a continual focus on addressing deficiencies.
arXiv Detail & Related papers (2022-08-19T00:55:05Z) - Fast nonlinear risk assessment for autonomous vehicles using learned
conditional probabilistic models of agent futures [19.247932561037487]
This paper presents fast non-sampling based methods to assess the risk for trajectories of autonomous vehicles.
The presented methods address a wide range of representations for uncertain predictions including both Gaussian and non-Gaussian mixture models.
We construct deterministic linear dynamical systems that govern the exact time evolution of the moments of uncertain position.
arXiv Detail & Related papers (2021-09-21T05:55:39Z) - Risk Minimization from Adaptively Collected Data: Guarantees for
Supervised and Policy Learning [57.88785630755165]
Empirical risk minimization (ERM) is the workhorse of machine learning, but its model-agnostic guarantees can fail when we use adaptively collected data.
We study a generic importance sampling weighted ERM algorithm for using adaptively collected data to minimize the average of a loss function over a hypothesis class.
For policy learning, we provide rate-optimal regret guarantees that close an open gap in the existing literature whenever exploration decays to zero.
arXiv Detail & Related papers (2021-06-03T09:50:13Z) - PAC$^m$-Bayes: Narrowing the Empirical Risk Gap in the Misspecified
Bayesian Regime [75.19403612525811]
This work develops a multi-sample loss which can close the gap by spanning a trade-off between the two risks.
Empirical study demonstrates improvement to the predictive distribution.
arXiv Detail & Related papers (2020-10-19T16:08:34Z) - Learning Bounds for Risk-sensitive Learning [86.50262971918276]
In risk-sensitive learning, one aims to find a hypothesis that minimizes a risk-averse (or risk-seeking) measure of loss.
We study the generalization properties of risk-sensitive learning schemes whose optimand is described via optimized certainty equivalents.
arXiv Detail & Related papers (2020-06-15T05:25:02Z) - Orthogonal Statistical Learning [49.55515683387805]
We provide non-asymptotic excess risk guarantees for statistical learning in a setting where the population risk depends on an unknown nuisance parameter.
We show that if the population risk satisfies a condition called Neymanity, the impact of the nuisance estimation error on the excess risk bound achieved by the meta-algorithm is of second order.
arXiv Detail & Related papers (2019-01-25T02:21:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.