A Statistical Learning Take on the Concordance Index for Survival
Analysis
- URL: http://arxiv.org/abs/2302.12059v1
- Date: Thu, 23 Feb 2023 14:33:54 GMT
- Title: A Statistical Learning Take on the Concordance Index for Survival
Analysis
- Authors: Alex Nowak-Vila, Kevin Elgui, Genevieve Robin
- Abstract summary: We provide C-index Fisher-consistency results and excess risk bounds for several commonly used cost functions in survival analysis.
We also study the general case where no model assumption is made and present a new, off-the-shelf method that is shown to be consistent with the C-index.
- Score: 0.29005223064604074
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The introduction of machine learning (ML) techniques to the field of survival
analysis has increased the flexibility of modeling approaches, and ML based
models have become state-of-the-art. These models optimize their own cost
functions, and their performance is often evaluated using the concordance index
(C-index). From a statistical learning perspective, it is therefore an
important problem to analyze the relationship between the optimizers of the
C-index and those of the ML cost functions. We address this issue by providing
C-index Fisher-consistency results and excess risk bounds for several of the
commonly used cost functions in survival analysis. We identify conditions under
which they are consistent, under the form of three nested families of survival
models. We also study the general case where no model assumption is made and
present a new, off-the-shelf method that is shown to be consistent with the
C-index, although computationally expensive at inference. Finally, we perform
limited numerical experiments with simulated data to illustrate our theoretical
findings.
Related papers
- FastSurvival: Hidden Computational Blessings in Training Cox Proportional Hazards Models [24.041562124587262]
Cox proportional hazards (CPH) model is widely used for interpretability, flexibility, and predictive performance.
Current algorithms to train the CPH model have drawbacks, preventing us from using the CPH model at its full potential.
We propose new optimization methods by constructing and minimizing surrogate functions that exploit hidden mathematical structures of the CPH model.
arXiv Detail & Related papers (2024-10-24T18:36:59Z) - On the KL-Divergence-based Robust Satisficing Model [2.425685918104288]
robustness satisficing framework has attracted increasing attention from academia.
We present analytical interpretations, diverse performance guarantees, efficient and stable numerical methods, convergence analysis, and an extension tailored for hierarchical data structures.
We demonstrate the superior performance of our model compared to state-of-the-art benchmarks.
arXiv Detail & Related papers (2024-08-17T10:05:05Z) - Learning Car-Following Behaviors Using Bayesian Matrix Normal Mixture Regression [17.828808886958736]
Car-following (CF) behaviors are crucial for microscopic traffic simulation.
Many data-driven methods, despite their robustness, operate as "black boxes" with limited interpretability.
This work introduces a Bayesian Matrix Normal Mixture Regression (MNMR) model that simultaneously captures feature correlations and temporal dynamics inherent in CF behaviors.
arXiv Detail & Related papers (2024-04-24T17:55:47Z) - Towards Stable Machine Learning Model Retraining via Slowly Varying Sequences [6.067007470552307]
We propose a methodology for finding sequences of machine learning models that are stable across retraining iterations.
We develop a mixed-integer optimization formulation that is guaranteed to recover optimal models.
Our method shows stronger stability than greedily trained models with a small, controllable sacrifice in predictive power.
arXiv Detail & Related papers (2024-03-28T22:45:38Z) - Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - The Common Stability Mechanism behind most Self-Supervised Learning
Approaches [64.40701218561921]
We provide a framework to explain the stability mechanism of different self-supervised learning techniques.
We discuss the working mechanism of contrastive techniques like SimCLR, non-contrastive techniques like BYOL, SWAV, SimSiam, Barlow Twins, and DINO.
We formulate different hypotheses and test them using the Imagenet100 dataset.
arXiv Detail & Related papers (2024-02-22T20:36:24Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - Explainable Censored Learning: Finding Critical Features with Long Term
Prognostic Values for Survival Prediction [28.943631598055926]
We introduce a novel, easily deployable approach, called EXplainable CEnsored Learning (EXCEL), to iteratively exploit critical variables.
We show that EXCEL can effectively identify critical features and achieve performance on par with or better than the original models.
arXiv Detail & Related papers (2022-09-30T12:56:29Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - Multiplicative noise and heavy tails in stochastic optimization [62.993432503309485]
empirical optimization is central to modern machine learning, but its role in its success is still unclear.
We show that it commonly arises in parameters of discrete multiplicative noise due to variance.
A detailed analysis is conducted in which we describe on key factors, including recent step size, and data, all exhibit similar results on state-of-the-art neural network models.
arXiv Detail & Related papers (2020-06-11T09:58:01Z) - Instability, Computational Efficiency and Statistical Accuracy [101.32305022521024]
We develop a framework that yields statistical accuracy based on interplay between the deterministic convergence rate of the algorithm at the population level, and its degree of (instability) when applied to an empirical object based on $n$ samples.
We provide applications of our general results to several concrete classes of models, including Gaussian mixture estimation, non-linear regression models, and informative non-response models.
arXiv Detail & Related papers (2020-05-22T22:30:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.