Individual dynamic prediction of clinical endpoint from large
dimensional longitudinal biomarker history: a landmark approach
- URL: http://arxiv.org/abs/2102.01466v1
- Date: Tue, 2 Feb 2021 12:36:18 GMT
- Title: Individual dynamic prediction of clinical endpoint from large
dimensional longitudinal biomarker history: a landmark approach
- Authors: Anthony Devaux (BPH), Robin Genuer (BPH, SISTM), Karine P\'er\`es
(BPH), C\'ecile Proust-Lima (BPH)
- Abstract summary: We propose a solution for the dynamic prediction of a health event that may exploit repeated measures of a possibly large number of markers.
Our methodology, implemented in R, enables the prediction of an event using the entire longitudinal patient history, even when the number of repeated markers is large.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The individual data collected throughout patient follow-up constitute crucial
information for assessing the risk of a clinical event, and eventually for
adapting a therapeutic strategy. Joint models and landmark models have been
proposed to compute individual dynamic predictions from repeated measures to
one or two markers. However, they hardly extend to the case where the complete
patient history includes much more repeated markers possibly. Our objective was
thus to propose a solution for the dynamic prediction of a health event that
may exploit repeated measures of a possibly large number of markers. We
combined a landmark approach extended to endogenous markers history with
machine learning methods adapted to survival data. Each marker trajectory is
modeled using the information collected up to landmark time, and summary
variables that best capture the individual trajectories are derived. These
summaries and additional covariates are then included in different prediction
methods. To handle a possibly large dimensional history, we rely on machine
learning methods adapted to survival data, namely regularized regressions and
random survival forests, to predict the event from the landmark time, and we
show how they can be combined into a superlearner. Then, the performances are
evaluated by cross-validation using estimators of Brier Score and the area
under the Receiver Operating Characteristic curve adapted to censored data. We
demonstrate in a simulation study the benefits of machine learning survival
methods over standard survival models, especially in the case of numerous
and/or nonlinear relationships between the predictors and the event. We then
applied the methodology in two prediction contexts: a clinical context with the
prediction of death for patients with primary biliary cholangitis, and a public
health context with the prediction of death in the general elderly population
at different ages. Our methodology, implemented in R, enables the prediction of
an event using the entire longitudinal patient history, even when the number of
repeated markers is large. Although introduced with mixed models for the
repeated markers and methods for a single right censored time-to-event, our
method can be used with any other appropriate modeling technique for the
markers and can be easily extended to competing risks setting.
Related papers
- SeqRisk: Transformer-augmented latent variable model for improved survival prediction with longitudinal data [4.1476925904032464]
We propose SeqRisk, a method that combines variational autoencoder (VAE) or longitudinal VAE (LVAE) with a transformer encoder and Cox proportional hazards module for risk prediction.
We demonstrate that SeqRisk performs competitively compared to existing approaches on both simulated and real-world datasets.
arXiv Detail & Related papers (2024-09-19T12:35:25Z) - Deep State-Space Generative Model For Correlated Time-to-Event Predictions [54.3637600983898]
We propose a deep latent state-space generative model to capture the interactions among different types of correlated clinical events.
Our method also uncovers meaningful insights about the latent correlations among mortality and different types of organ failures.
arXiv Detail & Related papers (2024-07-28T02:42:36Z) - Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - Kernel Cox partially linear regression: building predictive models for
cancer patients' survival [4.230753712933184]
We build a kernel Cox proportional hazards semi-parametric model and propose a novel regularized garrotized kernel machine (RegGKM) method to fit the model.
We use the kernel machine method to describe the complex relationship between survival and predictors, while automatically removing irrelevant parametric and non-parametric predictors.
Our results can help classify patients into groups with different death risks, facilitating treatment for better clinical outcomes.
arXiv Detail & Related papers (2023-10-11T04:27:54Z) - CenTime: Event-Conditional Modelling of Censoring in Survival Analysis [49.44664144472712]
We introduce CenTime, a novel approach to survival analysis that directly estimates the time to event.
Our method features an innovative event-conditional censoring mechanism that performs robustly even when uncensored data is scarce.
Our results indicate that CenTime offers state-of-the-art performance in predicting time-to-death while maintaining comparable ranking performance.
arXiv Detail & Related papers (2023-09-07T17:07:33Z) - Random survival forests for competing risks with multivariate
longitudinal endogenous covariates [0.0]
We propose an innovative solution to predict an event probability using a possibly large number of longitudinal predictors.
DynForest is an extension of random survival forests for competing risks that handles endogenous longitudinal predictors.
arXiv Detail & Related papers (2022-08-11T12:58:11Z) - Calibration of prediction rules for life-time outcomes using prognostic
Cox regression survival models and multiple imputations to account for
missing predictor data with cross-validatory assessment [0.0]
Methods are described to combine imputation with predictive calibration in survival modeling subject to censoring.
Prediction-averaging appears to have superior statistical properties, especially smaller predictive variation, as opposed to a direct application of Rubin's rules.
arXiv Detail & Related papers (2021-05-04T20:10:12Z) - Bootstrapping Your Own Positive Sample: Contrastive Learning With
Electronic Health Record Data [62.29031007761901]
This paper proposes a novel contrastive regularized clinical classification model.
We introduce two unique positive sampling strategies specifically tailored for EHR data.
Our framework yields highly competitive experimental results in predicting the mortality risk on real-world COVID-19 EHR data.
arXiv Detail & Related papers (2021-04-07T06:02:04Z) - Deep Cox Mixtures for Survival Regression [11.64579638651557]
We describe a new approach for survival analysis regression models, based on learning mixtures of Cox regressions to model individual survival distributions.
We perform experiments on multiple real world datasets, and look at the mortality rates of patients across ethnicity and gender.
arXiv Detail & Related papers (2021-01-16T22:41:22Z) - Two-step penalised logistic regression for multi-omic data with an
application to cardiometabolic syndrome [62.997667081978825]
We implement a two-step approach to multi-omic logistic regression in which variable selection is performed on each layer separately.
Our approach should be preferred if the goal is to select as many relevant predictors as possible.
Our proposed approach allows us to identify features that characterise cardiometabolic syndrome at the molecular level.
arXiv Detail & Related papers (2020-08-01T10:36:27Z) - Trajectories, bifurcations and pseudotime in large clinical datasets:
applications to myocardial infarction and diabetes data [94.37521840642141]
We suggest a semi-supervised methodology for the analysis of large clinical datasets, characterized by mixed data types and missing values.
The methodology is based on application of elastic principal graphs which can address simultaneously the tasks of dimensionality reduction, data visualization, clustering, feature selection and quantifying the geodesic distances (pseudotime) in partially ordered sequences of observations.
arXiv Detail & Related papers (2020-07-07T21:04:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.