Benchmarking state-of-the-art gradient boosting algorithms for
classification
- URL: http://arxiv.org/abs/2305.17094v1
- Date: Fri, 26 May 2023 17:06:15 GMT
- Title: Benchmarking state-of-the-art gradient boosting algorithms for
classification
- Authors: Piotr Florek, Adam Zagda\'nski
- Abstract summary: This work explores the use of gradient boosting in the context of classification.
Four popular implementations, including original GBM algorithm and selected state-of-the-art gradient boosting frameworks, have been compared.
An attempt was made to indicate a gradient boosting variant showing the right balance between effectiveness, reliability and ease of use.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work explores the use of gradient boosting in the context of
classification. Four popular implementations, including original GBM algorithm
and selected state-of-the-art gradient boosting frameworks (i.e. XGBoost,
LightGBM and CatBoost), have been thoroughly compared on several publicly
available real-world datasets of sufficient diversity. In the study, special
emphasis was placed on hyperparameter optimization, specifically comparing two
tuning strategies, i.e. randomized search and Bayesian optimization using the
Tree-stuctured Parzen Estimator. The performance of considered methods was
investigated in terms of common classification accuracy metrics as well as
runtime and tuning time. Additionally, obtained results have been validated
using appropriate statistical testing. An attempt was made to indicate a
gradient boosting variant showing the right balance between effectiveness,
reliability and ease of use.
Related papers
- OGBoost: A Python Package for Ordinal Gradient Boosting [0.0]
We introduce OGBoost, a scikit-learn-compatible Python package for ordinal regression using gradient boosting.
The package is available on PyPI and can be installed via "pip install ogboost"
arXiv Detail & Related papers (2025-02-19T06:06:12Z) - From Point to probabilistic gradient boosting for claim frequency and severity prediction [1.3812010983144802]
We present a unified notation, and contrast, all the existing point and probabilistic gradient boosting for decision tree algorithms.
We compare their performance on five publicly available datasets for claim frequency and severity, of various size and comprising different number of (high cardinality) categorical variables.
arXiv Detail & Related papers (2024-12-19T14:50:10Z) - Performance Embeddings: A Similarity-based Approach to Automatic
Performance Optimization [71.69092462147292]
Performance embeddings enable knowledge transfer of performance tuning between applications.
We demonstrate this transfer tuning approach on case studies in deep neural networks, dense and sparse linear algebra compositions, and numerical weather prediction stencils.
arXiv Detail & Related papers (2023-03-14T15:51:35Z) - An efficient hybrid classification approach for COVID-19 based on Harris
Hawks Optimization and Salp Swarm Optimization [0.0]
This study presents a hybrid binary version of the Harris Hawks Optimization algorithm (HHO) and Salp Swarm Optimization (SSA) for Covid-19 classification.
The proposed algorithm (HHOSSA) achieved 96% accuracy with the SVM, 98% and 98% accuracy with two classifiers.
arXiv Detail & Related papers (2022-12-25T19:52:18Z) - Dynamic Cat Swarm Optimization Algorithm for Backboard Wiring Problem [0.9990687944474739]
This paper presents a powerful swarm intelligence meta-heuristic optimization algorithm called Dynamic Cat Swarm Optimization.
The proposed algorithm suggests a new method to provide a proper balance between these phases by modifying the selection scheme and the seeking mode of the algorithm.
optimization results show the effectiveness of the proposed algorithm, which ranks first compared to several well-known algorithms available in the literature.
arXiv Detail & Related papers (2021-04-27T19:41:27Z) - Meta-Regularization: An Approach to Adaptive Choice of the Learning Rate
in Gradient Descent [20.47598828422897]
We propose textit-Meta-Regularization, a novel approach for the adaptive choice of the learning rate in first-order descent methods.
Our approach modifies the objective function by adding a regularization term, and casts the joint process parameters.
arXiv Detail & Related papers (2021-04-12T13:13:34Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Self-Tuning Stochastic Optimization with Curvature-Aware Gradient
Filtering [53.523517926927894]
We explore the use of exact per-sample Hessian-vector products and gradients to construct self-tuning quadratics.
We prove that our model-based procedure converges in noisy gradient setting.
This is an interesting step for constructing self-tuning quadratics.
arXiv Detail & Related papers (2020-11-09T22:07:30Z) - Pairwise Supervised Hashing with Bernoulli Variational Auto-Encoder and
Self-Control Gradient Estimator [62.26981903551382]
Variational auto-encoders (VAEs) with binary latent variables provide state-of-the-art performance in terms of precision for document retrieval.
We propose a pairwise loss function with discrete latent VAE to reward within-class similarity and between-class dissimilarity for supervised hashing.
This new semantic hashing framework achieves superior performance compared to the state-of-the-arts.
arXiv Detail & Related papers (2020-05-21T06:11:33Z) - Stochastic batch size for adaptive regularization in deep network
optimization [63.68104397173262]
We propose a first-order optimization algorithm incorporating adaptive regularization applicable to machine learning problems in deep learning framework.
We empirically demonstrate the effectiveness of our algorithm using an image classification task based on conventional network models applied to commonly used benchmark datasets.
arXiv Detail & Related papers (2020-04-14T07:54:53Z) - Variance Reduction with Sparse Gradients [82.41780420431205]
Variance reduction methods such as SVRG and SpiderBoost use a mixture of large and small batch gradients.
We introduce a new sparsity operator: The random-top-k operator.
Our algorithm consistently outperforms SpiderBoost on various tasks including image classification, natural language processing, and sparse matrix factorization.
arXiv Detail & Related papers (2020-01-27T08:23:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.