SurvNAM: The machine learning survival model explanation
- URL: http://arxiv.org/abs/2104.08903v1
- Date: Sun, 18 Apr 2021 16:40:56 GMT
- Title: SurvNAM: The machine learning survival model explanation
- Authors: Lev V. Utkin and Egor D. Satyukov and Andrei V. Konstantinov
- Abstract summary: SurvNAM is proposed to explain predictions of the black-box machine learning survival model.
The basic idea behind SurvNAM is to train the network by means of a specific expected loss function.
The proposed modifications of SurvNAM are based on using the Lasso-based regularization for functions from GAM.
- Score: 5.8010446129208155
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A new modification of the Neural Additive Model (NAM) called SurvNAM and its
modifications are proposed to explain predictions of the black-box machine
learning survival model. The method is based on applying the original NAM to
solving the explanation problem in the framework of survival analysis. The
basic idea behind SurvNAM is to train the network by means of a specific
expected loss function which takes into account peculiarities of the survival
model predictions and is based on approximating the black-box model by the
extension of the Cox proportional hazards model which uses the well-known
Generalized Additive Model (GAM) in place of the simple linear relationship of
covariates. The proposed method SurvNAM allows performing the local and global
explanation. A set of examples around the explained example is randomly
generated for the local explanation. The global explanation uses the whole
training dataset. The proposed modifications of SurvNAM are based on using the
Lasso-based regularization for functions from GAM and for a special
representation of the GAM functions using their weighted linear and non-linear
parts, which is implemented as a shortcut connection. A lot of numerical
experiments illustrate the SurvNAM efficiency.
Related papers
- Latent Space Energy-based Neural ODEs [73.01344439786524]
This paper introduces a novel family of deep dynamical models designed to represent continuous-time sequence data.
We train the model using maximum likelihood estimation with Markov chain Monte Carlo.
Experiments on oscillating systems, videos and real-world state sequences (MuJoCo) illustrate that ODEs with the learnable energy-based prior outperform existing counterparts.
arXiv Detail & Related papers (2024-09-05T18:14:22Z) - CoxSE: Exploring the Potential of Self-Explaining Neural Networks with Cox Proportional Hazards Model for Survival Analysis [3.977091269971331]
We explore the potential of self-explaining neural networks (SENN) for survival analysis.
We propose a new locally explainable Cox proportional hazards model, named CoxSE, by estimating a locally-linear log-hazard function.
We also propose a modification to the Neural additive (NAM) models hybrid with SENN, named CoxSENAM, which enables the control of the stability and consistency of the generated explanations.
arXiv Detail & Related papers (2024-07-18T18:32:54Z) - On the Origins of Linear Representations in Large Language Models [51.88404605700344]
We introduce a simple latent variable model to formalize the concept dynamics of the next token prediction.
Experiments show that linear representations emerge when learning from data matching the latent variable model.
We additionally confirm some predictions of the theory using the LLaMA-2 large language model.
arXiv Detail & Related papers (2024-03-06T17:17:36Z) - SurvBeNIM: The Beran-Based Neural Importance Model for Explaining the
Survival Models [2.4861619769660637]
A new method called the Survival Beran-based Neural Importance Model (SurvBeNIM) is proposed.
It aims to explain predictions of machine learning survival models, which are in the form of survival or cumulative hazard functions.
arXiv Detail & Related papers (2023-12-11T18:54:26Z) - On the Generalization and Adaption Performance of Causal Models [99.64022680811281]
Differentiable causal discovery has proposed to factorize the data generating process into a set of modules.
We study the generalization and adaption performance of such modular neural causal models.
Our analysis shows that the modular neural causal models outperform other models on both zero and few-shot adaptation in low data regimes.
arXiv Detail & Related papers (2022-06-09T17:12:32Z) - Sparse Neural Additive Model: Interpretable Deep Learning with Feature
Selection via Group Sparsity [22.752160137192156]
We study the theoretical properties for neural additive models (SNAM) with novel techniques to tackle the non-parametric truth.
We prove that SNAM can have exact support recovery, i.e. perfect feature selection, with appropriate regularization.
We further testify to the good accuracy and efficiency of SNAM.
arXiv Detail & Related papers (2022-02-25T03:40:53Z) - Leveraging Global Parameters for Flow-based Neural Posterior Estimation [90.21090932619695]
Inferring the parameters of a model based on experimental observations is central to the scientific method.
A particularly challenging setting is when the model is strongly indeterminate, i.e., when distinct sets of parameters yield identical observations.
We present a method for cracking such indeterminacy by exploiting additional information conveyed by an auxiliary set of observations sharing global parameters.
arXiv Detail & Related papers (2021-02-12T12:23:13Z) - Explanation-Guided Training for Cross-Domain Few-Shot Classification [96.12873073444091]
Cross-domain few-shot classification task (CD-FSC) combines few-shot classification with the requirement to generalize across domains represented by datasets.
We introduce a novel training approach for existing FSC models.
We show that explanation-guided training effectively improves the model generalization.
arXiv Detail & Related papers (2020-07-17T07:28:08Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - SurvLIME-Inf: A simplified modification of SurvLIME for explanation of
machine learning survival models [4.640835690336653]
The basic idea behind SurvLIME as well as SurvLIME-Inf is to apply the Cox proportional hazards model to approximate the black-box survival model at the local area around a test example.
In contrast to SurvLIME, the proposed modification uses $L_infty $-norm for defining distances between approximating and approximated cumulative hazard functions.
arXiv Detail & Related papers (2020-05-05T14:34:46Z) - SurvLIME: A method for explaining machine learning survival models [4.640835690336653]
The main idea behind the proposed method is to apply the Cox proportional hazards model to approximate the survival model at the local area around a test example.
A lot of numerical experiments demonstrate the SurvLIME efficiency.
arXiv Detail & Related papers (2020-03-18T17:48:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.