BoostTree and BoostForest for Ensemble Learning
- URL: http://arxiv.org/abs/2003.09737v3
- Date: Tue, 6 Dec 2022 06:32:49 GMT
- Title: BoostTree and BoostForest for Ensemble Learning
- Authors: Changming Zhao, Dongrui Wu, Jian Huang, Ye Yuan, Hai-Tao Zhang, Ruimin
Peng, Zhenhua Shi
- Abstract summary: BoostForest is an ensemble learning approach using BoostTree as base learners and can be used for both classification and regression.
It generally outperformed four classical ensemble learning approaches (Random Forest, Extra-Trees, XGBoost and LightGBM) on 35 classification and regression datasets.
- Score: 27.911350375268576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bootstrap aggregating (Bagging) and boosting are two popular ensemble
learning approaches, which combine multiple base learners to generate a
composite model for more accurate and more reliable performance. They have been
widely used in biology, engineering, healthcare, etc. This paper proposes
BoostForest, which is an ensemble learning approach using BoostTree as base
learners and can be used for both classification and regression. BoostTree
constructs a tree model by gradient boosting. It increases the randomness
(diversity) by drawing the cut-points randomly at node splitting. BoostForest
further increases the randomness by bootstrapping the training data in
constructing different BoostTrees. BoostForest generally outperformed four
classical ensemble learning approaches (Random Forest, Extra-Trees, XGBoost and
LightGBM) on 35 classification and regression datasets. Remarkably, BoostForest
tunes its parameters by simply sampling them randomly from a parameter pool,
which can be easily specified, and its ensemble learning framework can also be
used to combine many other base learners.
Related papers
- Binary Classification: Is Boosting stronger than Bagging? [5.877778007271621]
We introduce Enhanced Random Forests, an extension of vanilla Random Forests with extra functionalities and adaptive sample and model weighting.
We develop an iterative algorithm for adapting the training sample weights, by favoring the hardest examples, and an approach for finding personalized tree weighting schemes for each new sample.
Our method significantly improves upon regular Random Forests across 15 different binary classification datasets and considerably outperforms other tree methods, including XGBoost.
arXiv Detail & Related papers (2024-10-24T23:22:33Z) - Learning a Decision Tree Algorithm with Transformers [75.96920867382859]
We introduce MetaTree, a transformer-based model trained via meta-learning to directly produce strong decision trees.
We fit both greedy decision trees and globally optimized decision trees on a large number of datasets, and train MetaTree to produce only the trees that achieve strong generalization performance.
arXiv Detail & Related papers (2024-02-06T07:40:53Z) - A generalized decision tree ensemble based on the NeuralNetworks
architecture: Distributed Gradient Boosting Forest (DGBF) [0.0]
We present a graph-structured-tree-ensemble algorithm with a distributed representation learning process between trees naturally.
We call this novel approach Distributed Gradient Boosting Forest (DGBF) and we demonstrate that both RandomForest and GradientBoosting can be expressed as particular graph architectures of DGBF.
Finally, we see that the distributed learning outperforms both RandomForest and GradientBoosting in 7 out of 9 datasets.
arXiv Detail & Related papers (2024-02-04T09:22:52Z) - Multiclass Boosting: Simple and Intuitive Weak Learning Criteria [72.71096438538254]
We give a simple and efficient boosting algorithm, that does not require realizability assumptions.
We present a new result on boosting for list learners, as well as provide a novel proof for the characterization of multiclass PAC learning.
arXiv Detail & Related papers (2023-07-02T19:26:58Z) - SoftTreeMax: Policy Gradient with Tree Search [72.9513807133171]
We introduce SoftTreeMax, the first approach that integrates tree-search into policy gradient.
On Atari, SoftTreeMax demonstrates up to 5x better performance in faster run-time compared with distributed PPO.
arXiv Detail & Related papers (2022-09-28T09:55:47Z) - Fast ABC-Boost: A Unified Framework for Selecting the Base Class in
Multi-Class Classification [21.607059258448594]
We develop a unified framework for effectively selecting the base class by introducing a series of ideas to improve the computational efficiency of ABC-Boost.
Our framework has parameters $(s,g,w)$.
arXiv Detail & Related papers (2022-05-22T20:42:26Z) - To Boost or not to Boost: On the Limits of Boosted Neural Networks [67.67776094785363]
Boosting is a method for learning an ensemble of classifiers.
While boosting has been shown to be very effective for decision trees, its impact on neural networks has not been extensively studied.
We find that a single neural network usually generalizes better than a boosted ensemble of smaller neural networks with the same total number of parameters.
arXiv Detail & Related papers (2021-07-28T19:10:03Z) - Boost-R: Gradient Boosted Trees for Recurrence Data [13.40931458200203]
This paper investigates an additive-tree-based approach, known as Boost-R (Boosting for Recurrence Data), for recurrent event data with both static and dynamic features.
Boost-R constructs an ensemble of gradient boosted additive trees to estimate the cumulative intensity function of the recurrent event process.
arXiv Detail & Related papers (2021-07-03T02:44:09Z) - An Efficient Adversarial Attack for Tree Ensembles [91.05779257472675]
adversarial attacks on tree based ensembles such as gradient boosting decision trees (DTs) and random forests (RFs)
We show that our method can be thousands of times faster than the previous mixed-integer linear programming (MILP) based approach.
Our code is available at https://chong-z/tree-ensemble-attack.
arXiv Detail & Related papers (2020-10-22T10:59:49Z) - agtboost: Adaptive and Automatic Gradient Tree Boosting Computations [0.0]
agtboost implements fast gradient tree boosting computations.
A useful model validation function performs the Kolmogorov-Smirnov test on the learned distribution.
arXiv Detail & Related papers (2020-08-28T12:42:19Z) - Soft Gradient Boosting Machine [72.54062017726154]
We propose the soft Gradient Boosting Machine (sGBM) by wiring multiple differentiable base learners together.
Experimental results showed that, sGBM enjoys much higher time efficiency with better accuracy, given the same base learner in both on-line and off-line settings.
arXiv Detail & Related papers (2020-06-07T06:43:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.