Soft Gradient Boosting Machine
- URL: http://arxiv.org/abs/2006.04059v1
- Date: Sun, 7 Jun 2020 06:43:23 GMT
- Title: Soft Gradient Boosting Machine
- Authors: Ji Feng, Yi-Xuan Xu, Yuan Jiang, Zhi-Hua Zhou
- Abstract summary: We propose the soft Gradient Boosting Machine (sGBM) by wiring multiple differentiable base learners together.
Experimental results showed that, sGBM enjoys much higher time efficiency with better accuracy, given the same base learner in both on-line and off-line settings.
- Score: 72.54062017726154
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Gradient Boosting Machine has proven to be one successful function
approximator and has been widely used in a variety of areas. However, since the
training procedure of each base learner has to take the sequential order, it is
infeasible to parallelize the training process among base learners for
speed-up. In addition, under online or incremental learning settings, GBMs
achieved sub-optimal performance due to the fact that the previously trained
base learners can not adapt with the environment once trained. In this work, we
propose the soft Gradient Boosting Machine (sGBM) by wiring multiple
differentiable base learners together, by injecting both local and global
objectives inspired from gradient boosting, all base learners can then be
jointly optimized with linear speed-up. When using differentiable soft decision
trees as base learner, such device can be regarded as an alternative version of
the (hard) gradient boosting decision trees with extra benefits. Experimental
results showed that, sGBM enjoys much higher time efficiency with better
accuracy, given the same base learner in both on-line and off-line settings.
Related papers
- Two Optimizers Are Better Than One: LLM Catalyst Empowers Gradient-Based Optimization for Prompt Tuning [69.95292905263393]
We show that gradient-based optimization and large language models (MsLL) are complementary to each other, suggesting a collaborative optimization approach.
Our code is released at https://www.guozix.com/guozix/LLM-catalyst.
arXiv Detail & Related papers (2024-05-30T06:24:14Z) - ELRA: Exponential learning rate adaption gradient descent optimization
method [83.88591755871734]
We present a novel, fast (exponential rate), ab initio (hyper-free) gradient based adaption.
The main idea of the method is to adapt the $alpha by situational awareness.
It can be applied to problems of any dimensions n and scales only linearly.
arXiv Detail & Related papers (2023-09-12T14:36:13Z) - Bayesian Generational Population-Based Training [35.70338636901159]
Population-Based Training (PBT) has led to impressive performance in several large scale settings.
We introduce two new innovations in PBT-style methods.
We show that these innovations lead to large performance gains.
arXiv Detail & Related papers (2022-07-19T16:57:38Z) - BBTv2: Pure Black-Box Optimization Can Be Comparable to Gradient Descent
for Few-Shot Learning [83.26610968655815]
Black-Box Tuning is a derivative-free approach to optimize continuous prompt tokens prepended to the input of language models.
We present BBTv2, a pure black-box optimization approach that can drive language models to achieve comparable results to gradient-based optimization.
arXiv Detail & Related papers (2022-05-23T11:10:19Z) - Efficient Differentiable Simulation of Articulated Bodies [89.64118042429287]
We present a method for efficient differentiable simulation of articulated bodies.
This enables integration of articulated body dynamics into deep learning frameworks.
We show that reinforcement learning with articulated systems can be accelerated using gradients provided by our method.
arXiv Detail & Related papers (2021-09-16T04:48:13Z) - Analytically Tractable Bayesian Deep Q-Learning [0.0]
We adapt the temporal difference Q-learning framework to make it compatible with the tractable approximate Gaussian inference (TAGI)
We demonstrate that TAGI can reach a performance comparable to backpropagation-trained networks.
arXiv Detail & Related papers (2021-06-21T13:11:52Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Federated Transfer Learning with Dynamic Gradient Aggregation [27.42998421786922]
This paper introduces a Federated Learning (FL) simulation platform for Acoustic Model training.
The proposed FL platform can support different tasks based on the adopted modular design.
It is shown to outperform the golden standard of distributed training in both convergence speed and overall model performance.
arXiv Detail & Related papers (2020-08-06T04:29:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.