Ensembles of Double Random Forest
- URL: http://arxiv.org/abs/2111.02010v1
- Date: Wed, 3 Nov 2021 04:19:41 GMT
- Title: Ensembles of Double Random Forest
- Authors: M.A. Ganaie, M. Tanveer, P.N. Suganthan, V. Snasel
- Abstract summary: We propose two approaches for generating ensembles of double random forest.
In the first approach, we propose a rotation based ensemble of double random forest.
In the second approach, we propose oblique ensembles of double random forest.
- Score: 1.7205106391379026
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An ensemble of decision trees is known as Random Forest. As suggested by
Breiman, the strength of unstable learners and the diversity among them are the
ensemble models' core strength. In this paper, we propose two approaches for
generating ensembles of double random forest. In the first approach, we propose
a rotation based ensemble of double random forest. In rotation based double
random forests, transformation or rotation of the feature space is generated at
each node. At each node different random feature subspace is chosen for
evaluation, hence the transformation at each node is different. Different
transformations result in better diversity among the base learners and hence,
better generalization performance. With the double random forest as base
learner, the data at each node is transformed via two different transformations
namely, principal component analysis and linear discriminant analysis. In the
second approach, we propose oblique ensembles of double random forest. Decision
trees in random forest and double random forest are univariate, and this
results in the generation of axis parallel split which fails to capture the
geometric structure of the data. Also, the standard random forest may not grow
sufficiently large decision trees resulting in suboptimal performance. To
capture the geometric properties and to grow the decision trees of sufficient
depth, we propose oblique ensembles of double random forest. The oblique
ensembles of double random forest models are multivariate decision trees. At
each non-leaf node, multisurface proximal support vector machine generates the
optimal plane for better generalization performance. Also, different
regularization techniques (Tikhonov regularisation and axis-parallel split
regularisation) are employed for tackling the small sample size problems in the
decision trees of oblique ensembles of double random forest.
Related papers
- Adaptive Split Balancing for Optimal Random Forest [8.916614661563893]
We propose a new random forest algorithm that constructs the trees using a novel adaptive split-balancing method.
Our method achieves optimality in simple, smooth scenarios while adaptively learning the tree structure from the data.
arXiv Detail & Related papers (2024-02-17T09:10:40Z) - Learning a Decision Tree Algorithm with Transformers [75.96920867382859]
We introduce MetaTree, a transformer-based model trained via meta-learning to directly produce strong decision trees.
We fit both greedy decision trees and globally optimized decision trees on a large number of datasets, and train MetaTree to produce only the trees that achieve strong generalization performance.
arXiv Detail & Related papers (2024-02-06T07:40:53Z) - Contextual Decision Trees [62.997667081978825]
We propose a multi-armed contextual bandit recommendation framework for feature-based selection of a single shallow tree of the learned ensemble.
The trained system, which works on top of the Random Forest, dynamically identifies a base predictor that is responsible for providing the final output.
arXiv Detail & Related papers (2022-07-13T17:05:08Z) - Explaining random forest prediction through diverse rulesets [0.0]
Local Tree eXtractor (LTreeX) is able to explain the forest prediction for a given test instance with a few diverse rules.
We show that our proposed approach substantially outperforms other explainable methods in terms of predictive performance.
arXiv Detail & Related papers (2022-03-29T12:54:57Z) - Minimax Rates for High-Dimensional Random Tessellation Forests [0.0]
Mondrian forests is the first class of random forests for which minimax rates were obtained in arbitrary dimension.
We show that a large class of random forests with general split directions also achieve minimax optimal convergence rates in arbitrary dimension.
arXiv Detail & Related papers (2021-09-22T06:47:38Z) - Growing Deep Forests Efficiently with Soft Routing and Learned
Connectivity [79.83903179393164]
This paper further extends the deep forest idea in several important aspects.
We employ a probabilistic tree whose nodes make probabilistic routing decisions, a.k.a., soft routing, rather than hard binary decisions.
Experiments on the MNIST dataset demonstrate that our empowered deep forests can achieve better or comparable performance than [1],[3].
arXiv Detail & Related papers (2020-12-29T18:05:05Z) - Convex Polytope Trees [57.56078843831244]
convex polytope trees (CPT) are proposed to expand the family of decision trees by an interpretable generalization of their decision boundary.
We develop a greedy method to efficiently construct CPT and scalable end-to-end training algorithms for the tree parameters when the tree structure is given.
arXiv Detail & Related papers (2020-10-21T19:38:57Z) - Improved Weighted Random Forest for Classification Problems [3.42658286826597]
The key to make well-performing ensemble model is in the diversity of the base models.
We propose several algorithms that intend to modify the weighting strategy of regular random forest.
The proposed models are able to introduce significant improvements compared to regular random forest.
arXiv Detail & Related papers (2020-09-01T16:08:45Z) - Stochastic Optimization Forests [60.523606291705214]
We show how to train forest decision policies by growing trees that choose splits to directly optimize the downstream decision quality, rather than splitting to improve prediction accuracy as in the standard random forest algorithm.
We show that our approximate splitting criteria can reduce running time hundredfold, while achieving performance close to forest algorithms that exactly re-optimize for every candidate split.
arXiv Detail & Related papers (2020-08-17T16:56:06Z) - Generalized and Scalable Optimal Sparse Decision Trees [56.35541305670828]
We present techniques that produce optimal decision trees over a variety of objectives.
We also introduce a scalable algorithm that produces provably optimal results in the presence of continuous variables.
arXiv Detail & Related papers (2020-06-15T19:00:11Z) - Fr\'echet random forests for metric space valued regression with non
euclidean predictors [0.0]
We introduce Fr'echet trees and Fr'echet random forests, which allow to handle data for which input and output variables take values in general metric spaces.
A consistency theorem for Fr'echet regressogram predictor using data-driven partitions is given and applied to Fr'echet purely uniformly random trees.
arXiv Detail & Related papers (2019-06-04T22:07:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.