Variable selection for nonlinear Cox regression model via deep learning
- URL: http://arxiv.org/abs/2211.09287v1
- Date: Thu, 17 Nov 2022 01:17:54 GMT
- Title: Variable selection for nonlinear Cox regression model via deep learning
- Authors: Kexuan Li
- Abstract summary: We extend the recently developed deep learning-based variable selection model LassoNet to survival data.
We apply the proposed methodology to analyze a real data set on diffuse large B-cell lymphoma.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variable selection problem for the nonlinear Cox regression model is
considered. In survival analysis, one main objective is to identify the
covariates that are associated with the risk of experiencing the event of
interest. The Cox proportional hazard model is being used extensively in
survival analysis in studying the relationship between survival times and
covariates, where the model assumes that the covariate has a log-linear effect
on the hazard function. However, this linearity assumption may not be satisfied
in practice. In order to extract a representative subset of features, various
variable selection approaches have been proposed for survival data under the
linear Cox model. However, there exists little literature on variable selection
for the nonlinear Cox model. To break this gap, we extend the recently
developed deep learning-based variable selection model LassoNet to survival
data. Simulations are provided to demonstrate the validity and effectiveness of
the proposed method. Finally, we apply the proposed methodology to analyze a
real data set on diffuse large B-cell lymphoma.
Related papers
- Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - Factor-Augmented Regularized Model for Hazard Regression [1.8021287677546953]
We propose a new model, Factor-Augmented Regularized Model for Hazard Regression (FarmHazard), to perform model selection in high-dimensional data.
We prove model selection consistency and estimation consistency under mild conditions.
We also develop a factor-augmented variable screening procedure to deal with strong correlations in ultra-high dimensional problems.
arXiv Detail & Related papers (2022-10-03T16:35:33Z) - FastCPH: Efficient Survival Analysis for Neural Networks [57.03275837523063]
We propose FastCPH, a new method that runs in linear time and supports both the standard Breslow and Efron methods for tied events.
We also demonstrate the performance of FastCPH combined with LassoNet, a neural network that provides interpretability through feature sparsity.
arXiv Detail & Related papers (2022-08-21T03:35:29Z) - X-model: Improving Data Efficiency in Deep Learning with A Minimax Model [78.55482897452417]
We aim at improving data efficiency for both classification and regression setups in deep learning.
To take the power of both worlds, we propose a novel X-model.
X-model plays a minimax game between the feature extractor and task-specific heads.
arXiv Detail & Related papers (2021-10-09T13:56:48Z) - Fast Bayesian Variable Selection in Binomial and Negative Binomial
Regression [9.774282306558465]
We introduce an efficient MCMC scheme for variable selection in binomial and negative binomial regression, that exploits logistic regression as a special case.
In experiments we demonstrate the effectiveness of our approach, including on data with seventeen thousand covariates.
arXiv Detail & Related papers (2021-06-28T20:54:41Z) - Variable selection with missing data in both covariates and outcomes:
Imputation and machine learning [1.0333430439241666]
The missing data issue is ubiquitous in health studies.
Machine learning methods weaken parametric assumptions.
XGBoost and BART have the overall best performance across various settings.
arXiv Detail & Related papers (2021-04-06T20:18:29Z) - Increasing the efficiency of randomized trial estimates via linear
adjustment for a prognostic score [59.75318183140857]
Estimating causal effects from randomized experiments is central to clinical research.
Most methods for historical borrowing achieve reductions in variance by sacrificing strict type-I error rate control.
arXiv Detail & Related papers (2020-12-17T21:10:10Z) - The Risks of Invariant Risk Minimization [52.7137956951533]
Invariant Risk Minimization is an objective based on the idea for learning deep, invariant features of data.
We present the first analysis of classification under the IRM objective--as well as these recently proposed alternatives--under a fairly natural and general model.
We show that IRM can fail catastrophically unless the test data are sufficiently similar to the training distribution--this is precisely the issue that it was intended to solve.
arXiv Detail & Related papers (2020-10-12T14:54:32Z) - DeepHazard: neural network for time-varying risks [0.6091702876917281]
We propose a new flexible method for survival prediction: DeepHazard, a neural network for time-varying risks.
Our approach is tailored for a wide range of continuous hazards forms, with the only restriction of being additive in time.
Numerical examples illustrate that our approach outperforms existing state-of-the-art methodology in terms of predictive capability evaluated through the C-index metric.
arXiv Detail & Related papers (2020-07-26T21:01:49Z) - A General Framework for Survival Analysis and Multi-State Modelling [70.31153478610229]
We use neural ordinary differential equations as a flexible and general method for estimating multi-state survival models.
We show that our model exhibits state-of-the-art performance on popular survival data sets and demonstrate its efficacy in a multi-state setting.
arXiv Detail & Related papers (2020-06-08T19:24:54Z) - SurvLIME: A method for explaining machine learning survival models [4.640835690336653]
The main idea behind the proposed method is to apply the Cox proportional hazards model to approximate the survival model at the local area around a test example.
A lot of numerical experiments demonstrate the SurvLIME efficiency.
arXiv Detail & Related papers (2020-03-18T17:48:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.