Optimal survival trees ensemble
- URL: http://arxiv.org/abs/2005.09043v1
- Date: Mon, 18 May 2020 19:28:16 GMT
- Title: Optimal survival trees ensemble
- Authors: Naz Gul, Nosheen Faiz, Dan Brawn, Rafal Kulakowski, Zardad Khan and
Berthold Lausen
- Abstract summary: Recent studies have adopted an approach of selecting accurate and diverse trees based on individual or collective performance within an ensemble for classification and regression problems.
This work follows in the wake of these investigations and considers the possibility of growing a forest of optimal survival trees.
In addition to improve predictive performance, the proposed method reduces the number of survival trees in the ensemble as compared to the other tree based methods.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have adopted an approach of selecting accurate and diverse
trees based on individual or collective performance within an ensemble for
classification and regression problems. This work follows in the wake of these
investigations and considers the possibility of growing a forest of optimal
survival trees. Initially, a large set of survival trees are grown using the
method of random survival forest. The grown trees are then ranked from smallest
to highest value of their prediction error using out-of-bag observations for
each respective survival tree. The top ranked survival trees are then assessed
for their collective performance as an ensemble. This ensemble is initiated
with the survival tree which stands first in rank, then further trees are
tested one by one by adding them to the ensemble in order of rank. A survival
tree is selected for the resultant ensemble if the performance improves after
an assessment using independent training data. This ensemble is called an
optimal trees ensemble (OSTE). The proposed method is assessed using 17
benchmark datasets and the results are compared with those of random survival
forest, conditional inference forest, bagging and a non tree based method, the
Cox proportional hazard model. In addition to improve predictive performance,
the proposed method reduces the number of survival trees in the ensemble as
compared to the other tree based methods. The method is implemented in an R
package called "OSTE".
Related papers
- Can a Single Tree Outperform an Entire Forest? [5.448070998907116]
The prevailing mindset is that a single decision tree underperforms classic random forests in testing accuracy.
This study challenges such a mindset by significantly improving the testing accuracy of an oblique regression tree.
Our approach reformulates tree training as a differentiable unconstrained optimization task.
arXiv Detail & Related papers (2024-11-26T00:18:18Z) - Heterogeneous Random Forest [2.0646127669654835]
Heterogeneous Random Forest (HRF) is designed to enhance tree diversity in a meaningful way.
HRF consistently outperformed other ensemble methods in terms of accuracy across the majority of datasets.
arXiv Detail & Related papers (2024-10-24T09:18:55Z) - Learning Deep Tree-based Retriever for Efficient Recommendation: Theory and Method [76.31185707649227]
We propose a Deep Tree-based Retriever (DTR) for efficient recommendation.
DTR frames the training task as a softmax-based multi-class classification over tree nodes at the same level.
To mitigate the suboptimality induced by the labeling of non-leaf nodes, we propose a rectification method for the loss function.
arXiv Detail & Related papers (2024-08-21T05:09:53Z) - Ensembles of Probabilistic Regression Trees [46.53457774230618]
Tree-based ensemble methods have been successfully used for regression problems in many applications and research studies.
We study ensemble versions of probabilisticregression trees that provide smooth approximations of the objective function by assigningeach observation to each region with respect to a probability distribution.
arXiv Detail & Related papers (2024-06-20T06:51:51Z) - Learning a Decision Tree Algorithm with Transformers [75.96920867382859]
We introduce MetaTree, a transformer-based model trained via meta-learning to directly produce strong decision trees.
We fit both greedy decision trees and globally optimized decision trees on a large number of datasets, and train MetaTree to produce only the trees that achieve strong generalization performance.
arXiv Detail & Related papers (2024-02-06T07:40:53Z) - Why do Random Forests Work? Understanding Tree Ensembles as
Self-Regularizing Adaptive Smoothers [68.76846801719095]
We argue that the current high-level dichotomy into bias- and variance-reduction prevalent in statistics is insufficient to understand tree ensembles.
We show that forests can improve upon trees by three distinct mechanisms that are usually implicitly entangled.
arXiv Detail & Related papers (2024-02-02T15:36:43Z) - Contextual Decision Trees [62.997667081978825]
We propose a multi-armed contextual bandit recommendation framework for feature-based selection of a single shallow tree of the learned ensemble.
The trained system, which works on top of the Random Forest, dynamically identifies a base predictor that is responsible for providing the final output.
arXiv Detail & Related papers (2022-07-13T17:05:08Z) - Social Interpretable Tree for Pedestrian Trajectory Prediction [75.81745697967608]
We propose a tree-based method, termed as Social Interpretable Tree (SIT), to address this multi-modal prediction task.
A path in the tree from the root to leaf represents an individual possible future trajectory.
Despite the hand-crafted tree, the experimental results on ETH-UCY and Stanford Drone datasets demonstrate that our method is capable of matching or exceeding the performance of state-of-the-art methods.
arXiv Detail & Related papers (2022-05-26T12:18:44Z) - Explaining random forest prediction through diverse rulesets [0.0]
Local Tree eXtractor (LTreeX) is able to explain the forest prediction for a given test instance with a few diverse rules.
We show that our proposed approach substantially outperforms other explainable methods in terms of predictive performance.
arXiv Detail & Related papers (2022-03-29T12:54:57Z) - Optimal trees selection for classification via out-of-bag assessment and
sub-bagging [0.0]
The predictive performance of tree based machine learning methods, in general, improves with a decreasing rate as the size of training data increases.
We investigate this in optimal trees ensemble (OTE) where the method fails to learn from some of the training observations due to internal validation.
Modified tree selection methods are thus proposed for OTE to cater for the loss of training observations in internal validation.
arXiv Detail & Related papers (2020-12-30T19:44:11Z) - Optimal Survival Trees [2.7910505923792637]
We present a new Optimal Survival Trees algorithm that leverages mixed-integer optimization (MIO) and local search techniques to generate globally optimized survival tree models.
We demonstrate that the algorithm improves on the accuracy of existing survival tree methods, particularly in large datasets.
arXiv Detail & Related papers (2020-12-08T09:00:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.