A Generalized Stacking for Implementing Ensembles of Gradient Boosting
Machines
- URL: http://arxiv.org/abs/2010.06026v1
- Date: Mon, 12 Oct 2020 21:05:45 GMT
- Title: A Generalized Stacking for Implementing Ensembles of Gradient Boosting
Machines
- Authors: Andrei V. Konstantinov and Lev V. Utkin
- Abstract summary: An approach for constructing ensembles of gradient boosting models is proposed.
It is shown that the proposed approach can be simply extended on arbitrary differentiable combination models.
- Score: 5.482532589225552
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The gradient boosting machine is one of the powerful tools for solving
regression problems. In order to cope with its shortcomings, an approach for
constructing ensembles of gradient boosting models is proposed. The main idea
behind the approach is to use the stacking algorithm in order to learn a
second-level meta-model which can be regarded as a model for implementing
various ensembles of gradient boosting models. First, the linear regression of
the gradient boosting models is considered as a simplest realization of the
meta-model under condition that the linear model is differentiable with respect
to its coefficients (weights). Then it is shown that the proposed approach can
be simply extended on arbitrary differentiable combination models, for example,
on neural networks which are differentiable and can implement arbitrary
functions of gradient boosting models. Various numerical examples illustrate
the proposed approach.
Related papers
- Supervised Score-Based Modeling by Gradient Boosting [49.556736252628745]
We propose a Supervised Score-based Model (SSM) which can be viewed as a gradient boosting algorithm combining score matching.
We provide a theoretical analysis of learning and sampling for SSM to balance inference time and prediction accuracy.
Our model outperforms existing models in both accuracy and inference time.
arXiv Detail & Related papers (2024-11-02T07:06:53Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - Geometrically Guided Integrated Gradients [0.3867363075280543]
We introduce an interpretability method called "geometrically-guided integrated gradients"
Our method explores the model's dynamic behavior from multiple scaled versions of the input and captures the best possible attribution for each input.
We also propose a "model perturbation" sanity check to complement the traditionally used "model randomization" test.
arXiv Detail & Related papers (2022-06-13T05:05:43Z) - Differentiable Spline Approximations [48.10988598845873]
Differentiable programming has significantly enhanced the scope of machine learning.
Standard differentiable programming methods (such as autodiff) typically require that the machine learning models be differentiable.
We show that leveraging this redesigned Jacobian in the form of a differentiable "layer" in predictive models leads to improved performance in diverse applications.
arXiv Detail & Related papers (2021-10-04T16:04:46Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z) - A Multilevel Approach to Training [0.0]
We propose a novel training method based on nonlinear multilevel techniques, commonly used for solving discretized large scale partial differential equations.
Our multilevel training method constructs a multilevel hierarchy by reducing the number of samples.
The training of the original model is then enhanced by internally training surrogate models constructed with fewer samples.
arXiv Detail & Related papers (2020-06-28T13:34:48Z) - Uncertainty in Gradient Boosting via Ensembles [37.808845398471874]
ensembles of gradient boosting models successfully detect anomalous inputs while having limited ability to improve the predicted total uncertainty.
We propose a concept of a virtual ensemble to get the benefits of an ensemble via only one gradient boosting model, which significantly reduces complexity.
arXiv Detail & Related papers (2020-06-18T14:11:27Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.