Regression Trees for Cumulative Incidence Functions
- URL: http://arxiv.org/abs/2011.06706v1
- Date: Fri, 13 Nov 2020 00:37:12 GMT
- Title: Regression Trees for Cumulative Incidence Functions
- Authors: Youngjoo Cho, Annette M. Molinaro, Chen Hu, and Robert L. Strawderman
- Abstract summary: We develop a novel approach to building regression trees for estimating cumulative incidence curves.
The proposed methods are easily implemented using the R statistical software package.
- Score: 3.0798859462300756
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The use of cumulative incidence functions for characterizing the risk of one
type of event in the presence of others has become increasingly popular over
the past decade. The problems of modeling, estimation and inference have been
treated using parametric, nonparametric and semi-parametric methods. Efforts to
develop suitable extensions of machine learning methods, such as regression
trees and related ensemble methods, have begun only recently. In this paper, we
develop a novel approach to building regression trees for estimating cumulative
incidence curves in a competing risks setting. The proposed methods employ
augmented estimators of the Brier score risk as the primary basis for building
and pruning trees. The proposed methods are easily implemented using the R
statistical software package. Simulation studies demonstrate the utility of our
approach in the competing risks setting. Data from the Radiation Therapy
Oncology Group (trial 9410) is used to illustrate these new methods.
Related papers
- FPBoost: Fully Parametric Gradient Boosting for Survival Analysis [4.09225917049674]
We propose a novel paradigm for survival model design based on the weighted sum of individual fully parametric hazard contributions.
The proposed model, which we call FPBoost, is the first algorithm to directly optimize the survival likelihood via gradient boosting.
arXiv Detail & Related papers (2024-09-20T09:57:17Z) - Regression Trees Know Calculus [1.8130068086063336]
We find a simple estimate of the gradient which can be efficiently computed using quantities exposed by popular tree learning libraries.
This allows the tools developed in the context of differentiable algorithms, like neural nets and Gaussian processes, to be deployed to tree-based models.
arXiv Detail & Related papers (2024-05-22T17:14:03Z) - Provable Risk-Sensitive Distributional Reinforcement Learning with
General Function Approximation [54.61816424792866]
We introduce a general framework on Risk-Sensitive Distributional Reinforcement Learning (RS-DisRL), with static Lipschitz Risk Measures (LRM) and general function approximation.
We design two innovative meta-algorithms: textttRS-DisRL-M, a model-based strategy for model-based function approximation, and textttRS-DisRL-V, a model-free approach for general value function approximation.
arXiv Detail & Related papers (2024-02-28T08:43:18Z) - Vector-Valued Least-Squares Regression under Output Regularity
Assumptions [73.99064151691597]
We propose and analyse a reduced-rank method for solving least-squares regression problems with infinite dimensional output.
We derive learning bounds for our method, and study under which setting statistical performance is improved in comparison to full-rank method.
arXiv Detail & Related papers (2022-11-16T15:07:00Z) - Robust online joint state/input/parameter estimation of linear systems [0.0]
This paper presents a method for jointly estimating the state, input, and parameters of linear systems in an online fashion.
The method is specially designed for measurements that are corrupted with non-Gaussian noise or outliers.
arXiv Detail & Related papers (2022-04-12T09:41:28Z) - Statistical Inference for High-Dimensional Linear Regression with
Blockwise Missing Data [13.48481978963297]
Blockwise missing data occurs when we integrate multisource or multimodality data where different sources or modalities contain complementary information.
We propose a computationally efficient estimator for the regression coefficient vector based on carefully constructed unbiased estimating equations.
Numerical studies and application analysis of the Alzheimer's Disease Neuroimaging Initiative data show that the proposed method performs better and benefits more from unsupervised samples than existing methods.
arXiv Detail & Related papers (2021-06-07T05:12:42Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Variable selection with missing data in both covariates and outcomes:
Imputation and machine learning [1.0333430439241666]
The missing data issue is ubiquitous in health studies.
Machine learning methods weaken parametric assumptions.
XGBoost and BART have the overall best performance across various settings.
arXiv Detail & Related papers (2021-04-06T20:18:29Z) - SAMBA: Safe Model-Based & Active Reinforcement Learning [59.01424351231993]
SAMBA is a framework for safe reinforcement learning that combines aspects from probabilistic modelling, information theory, and statistics.
We evaluate our algorithm on a variety of safe dynamical system benchmarks involving both low and high-dimensional state representations.
We provide intuition as to the effectiveness of the framework by a detailed analysis of our active metrics and safety constraints.
arXiv Detail & Related papers (2020-06-12T10:40:46Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z) - Adaptive Correlated Monte Carlo for Contextual Categorical Sequence
Generation [77.7420231319632]
We adapt contextual generation of categorical sequences to a policy gradient estimator, which evaluates a set of correlated Monte Carlo (MC) rollouts for variance control.
We also demonstrate the use of correlated MC rollouts for binary-tree softmax models, which reduce the high generation cost in large vocabulary scenarios.
arXiv Detail & Related papers (2019-12-31T03:01:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.