Factor-Augmented Regularized Model for Hazard Regression
- URL: http://arxiv.org/abs/2210.01067v1
- Date: Mon, 3 Oct 2022 16:35:33 GMT
- Title: Factor-Augmented Regularized Model for Hazard Regression
- Authors: Pierre Bayle, Jianqing Fan
- Abstract summary: We propose a new model, Factor-Augmented Regularized Model for Hazard Regression (FarmHazard), to perform model selection in high-dimensional data.
We prove model selection consistency and estimation consistency under mild conditions.
We also develop a factor-augmented variable screening procedure to deal with strong correlations in ultra-high dimensional problems.
- Score: 1.8021287677546953
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A prevalent feature of high-dimensional data is the dependence among
covariates, and model selection is known to be challenging when covariates are
highly correlated. To perform model selection for the high-dimensional Cox
proportional hazards model in presence of correlated covariates with factor
structure, we propose a new model, Factor-Augmented Regularized Model for
Hazard Regression (FarmHazard), which builds upon latent factors that drive
covariate dependence and extends Cox's model. This new model generates
procedures that operate in two steps by learning factors and idiosyncratic
components from high-dimensional covariate vectors and then using them as new
predictors. Cox's model is a widely used semi-parametric model for survival
analysis, where censored data and time-dependent covariates bring additional
technical challenges. We prove model selection consistency and estimation
consistency under mild conditions. We also develop a factor-augmented variable
screening procedure to deal with strong correlations in ultra-high dimensional
problems. Extensive simulations and real data experiments demonstrate that our
procedures enjoy good performance and achieve better results on model
selection, out-of-sample C-index and screening than alternative methods.
Related papers
- Less is More: Mitigate Spurious Correlations for Open-Domain Dialogue
Response Generation Models by Causal Discovery [52.95935278819512]
We conduct the first study on spurious correlations for open-domain response generation models based on a corpus CGDIALOG curated in our work.
Inspired by causal discovery algorithms, we propose a novel model-agnostic method for training and inference of response generation model.
arXiv Detail & Related papers (2023-03-02T06:33:48Z) - Strong identifiability and parameter learning in regression with
heterogeneous response [5.503319042839695]
We investigate conditions of strong identifiability, rates of convergence for conditional density and parameter estimation, and the Bayesian posterior contraction behavior arising in finite mixture of regression models.
We provide simulation studies and data illustrations, which shed some light on the parameter learning behavior found in several popular regression mixture models reported in the literature.
arXiv Detail & Related papers (2022-12-08T05:58:13Z) - Variable selection for nonlinear Cox regression model via deep learning [0.0]
We extend the recently developed deep learning-based variable selection model LassoNet to survival data.
We apply the proposed methodology to analyze a real data set on diffuse large B-cell lymphoma.
arXiv Detail & Related papers (2022-11-17T01:17:54Z) - Hybrid Feature- and Similarity-Based Models for Prediction and
Interpretation using Large-Scale Observational Data [0.0]
We propose a hybrid feature- and similarity-based model for supervised learning.
The proposed hybrid model is fit by convex optimization with a sparsity-inducing penalty on the kernel portion.
We compared our models to solely feature- and similarity-based approaches using synthetic data and using EHR data to predict risk of loneliness or social isolation.
arXiv Detail & Related papers (2022-04-12T20:37:03Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - A non-asymptotic model selection in block-diagonal mixture of polynomial
experts models [1.491109220586182]
We introduce a penalized maximum likelihood selection criterion to estimate the unknown conditional density of a regression model.
We provide a strong theoretical guarantee, including a finite-sample oracle satisfied by the penalized maximum likelihood with a Jensen-Kullback-Leibler type loss.
arXiv Detail & Related papers (2021-04-18T21:32:20Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - Autoregressive Score Matching [113.4502004812927]
We propose autoregressive conditional score models (AR-CSM) where we parameterize the joint distribution in terms of the derivatives of univariable log-conditionals (scores)
For AR-CSM models, this divergence between data and model distributions can be computed and optimized efficiently, requiring no expensive sampling or adversarial training.
We show with extensive experimental results that it can be applied to density estimation on synthetic data, image generation, image denoising, and training latent variable models with implicit encoders.
arXiv Detail & Related papers (2020-10-24T07:01:24Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z) - A General Framework for Survival Analysis and Multi-State Modelling [70.31153478610229]
We use neural ordinary differential equations as a flexible and general method for estimating multi-state survival models.
We show that our model exhibits state-of-the-art performance on popular survival data sets and demonstrate its efficacy in a multi-state setting.
arXiv Detail & Related papers (2020-06-08T19:24:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.