On the calibration of survival models with competing risks
- URL: http://arxiv.org/abs/2602.00194v1
- Date: Fri, 30 Jan 2026 10:23:33 GMT
- Title: On the calibration of survival models with competing risks
- Authors: Julie Alberge, Tristan Haugomat, Gaël Varoquaux, Judith Abécassis,
- Abstract summary: We show that existing calibration measures are not suited to the competing-risk setting.<n>We introduce two novel calibration measures that are minimized for oracle estimators.<n>Our recalibration methods yield good probabilities while preserving discrimination.
- Score: 9.77401900270683
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Survival analysis deals with modeling the time until an event occurs, and accurate probability estimates are crucial for decision-making, particularly in the competing-risks setting where multiple events are possible. While recent work has addressed calibration in standard survival analysis, the competing-risks setting remains under-explored as it is harder (the calibration applies to both probabilities across classes and time horizon). We show that existing calibration measures are not suited to the competing-risk setting and that recent models do not give well-behaved probabilities. To address this, we introduce a dedicated framework with two novel calibration measures that are minimized for oracle estimators (i.e., both measures are proper). We also introduce some methods to estimate, test, and correct the calibration. Our recalibration methods yield good probabilities while preserving discrimination.
Related papers
- Rethinking Early Stopping: Refine, Then Calibrate [49.966899634962374]
We present a novel variational formulation of the calibration-refinement decomposition.<n>We provide theoretical and empirical evidence that calibration and refinement errors are not minimized simultaneously during training.
arXiv Detail & Related papers (2025-01-31T15:03:54Z) - Enhancing accuracy of uncertainty estimation in appearance-based gaze tracking with probabilistic evaluation and calibration [13.564919425738163]
Uncertainty in appearance-based gaze tracking is critical for ensuring reliable downstream applications.<n>Current uncertainty-aware approaches adopt probabilistic models to acquire uncertainties by following distributions in the training dataset.<n>We propose a correction strategy based on probability calibration to mitigate biases in the estimated uncertainties of the trained models.
arXiv Detail & Related papers (2025-01-24T19:33:55Z) - Survival Models: Proper Scoring Rule and Stochastic Optimization with Competing Risks [6.9648613217501705]
SurvivalBoost outperforms 12 state-of-the-art models on 4 real-life datasets.
It provides great calibration, the ability to predict across any time horizon, and computation times faster than existing methods.
arXiv Detail & Related papers (2024-10-22T07:33:34Z) - Cautious Calibration in Binary Classification [1.6954767541769011]
Being cautious is crucial for enhancing the trustworthiness of machine learning systems.
In high-risk scenarios, it is important for each predicted probability to lean towards underconfidence.
In this study, we introduce the novel concept of cautious calibration in binary classification.
arXiv Detail & Related papers (2024-08-09T15:19:40Z) - Towards Certification of Uncertainty Calibration under Adversarial Attacks [96.48317453951418]
We show that attacks can significantly harm calibration, and thus propose certified calibration as worst-case bounds on calibration under adversarial perturbations.<n>We propose novel calibration attacks and demonstrate how they can improve model calibration through textitadversarial calibration training
arXiv Detail & Related papers (2024-05-22T18:52:09Z) - Calibration by Distribution Matching: Trainable Kernel Calibration
Metrics [56.629245030893685]
We introduce kernel-based calibration metrics that unify and generalize popular forms of calibration for both classification and regression.
These metrics admit differentiable sample estimates, making it easy to incorporate a calibration objective into empirical risk minimization.
We provide intuitive mechanisms to tailor calibration metrics to a decision task, and enforce accurate loss estimation and no regret decisions.
arXiv Detail & Related papers (2023-10-31T06:19:40Z) - Boldness-Recalibration for Binary Event Predictions [0.0]
Ideally, probability predictions are (i) well calibrated, (ii) accurate, and (iii) bold, i.e., spread out enough to be informative for decision making.
There is a fundamental tension between calibration and boldness, since calibration metrics can be high when predictions are overly cautious, i.e., non-bold.
The purpose of this work is to develop a Bayesian model selection-based approach to assess calibration, and a strategy for boldness-recalibration.
arXiv Detail & Related papers (2023-05-05T18:14:47Z) - Continuous Risk Measures for Driving Support [0.0]
We compare three model-based risk measures by evaluating their stengths and qualitatively testing them quantitatively.
We derive a novel risk measure based on the statistics of sparse critical events and so-called survival conditions.
The resulting survival analysis shows to have an earlier detection time crashes and less false positive detections in near-crash and non-crash cases supported by its solid theoretical grounding.
arXiv Detail & Related papers (2023-03-14T15:54:37Z) - Mitigating multiple descents: A model-agnostic framework for risk
monotonization [84.6382406922369]
We develop a general framework for risk monotonization based on cross-validation.
We propose two data-driven methodologies, namely zero- and one-step, that are akin to bagging and boosting.
arXiv Detail & Related papers (2022-05-25T17:41:40Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - Unsupervised Calibration under Covariate Shift [92.02278658443166]
We introduce the problem of calibration under domain shift and propose an importance sampling based approach to address it.
We evaluate and discuss the efficacy of our method on both real-world datasets and synthetic datasets.
arXiv Detail & Related papers (2020-06-29T21:50:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.