(Decision and regression) tree ensemble based kernels for regression and
classification
- URL: http://arxiv.org/abs/2012.10737v1
- Date: Sat, 19 Dec 2020 16:52:58 GMT
- Title: (Decision and regression) tree ensemble based kernels for regression and
classification
- Authors: Dai Feng and Richard Baumgartner
- Abstract summary: Tree based ensembles such as Breiman's random forest (RF) and Gradient Boosted Trees (GBT) can be interpreted as implicit kernel generators.
We show that for continuous targets, the RF/GBT kernels are competitive to their respective ensembles in higher dimensional scenarios.
We provide the results from real life data sets for regression and classification to show how these insights may be leveraged in practice.
- Score: 2.28438857884398
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Tree based ensembles such as Breiman's random forest (RF) and Gradient
Boosted Trees (GBT) can be interpreted as implicit kernel generators, where the
ensuing proximity matrix represents the data-driven tree ensemble kernel.
Kernel perspective on the RF has been used to develop a principled framework
for theoretical investigation of its statistical properties. Recently, it has
been shown that the kernel interpretation is germane to other tree-based
ensembles e.g. GBTs. However, practical utility of the links between kernels
and the tree ensembles has not been widely explored and systematically
evaluated.
Focus of our work is investigation of the interplay between kernel methods
and the tree based ensembles including the RF and GBT. We elucidate the
performance and properties of the RF and GBT based kernels in a comprehensive
simulation study comprising of continuous and binary targets. We show that for
continuous targets, the RF/GBT kernels are competitive to their respective
ensembles in higher dimensional scenarios, particularly in cases with larger
number of noisy features. For the binary target, the RF/GBT kernels and their
respective ensembles exhibit comparable performance. We provide the results
from real life data sets for regression and classification to show how these
insights may be leveraged in practice. Overall, our results support the tree
ensemble based kernels as a valuable addition to the practitioner's toolbox.
Finally, we discuss extensions of the tree ensemble based kernels for
survival targets, interpretable prototype and landmarking classification and
regression. We outline future line of research for kernels furnished by
Bayesian counterparts of the frequentist tree ensembles.
Related papers
- Parallel Tree Kernel Computation [0.0]
We devise a parallel implementation of the sequential algorithm for the computation of some tree kernels of two finite sets of trees.
Results show that the proposed parallel algorithm outperforms the sequential one in terms of latency.
arXiv Detail & Related papers (2023-05-12T18:16:45Z) - Meta-Learning Hypothesis Spaces for Sequential Decision-making [79.73213540203389]
We propose to meta-learn a kernel from offline data (Meta-KeL)
Under mild conditions, we guarantee that our estimated RKHS yields valid confidence sets.
We also empirically evaluate the effectiveness of our approach on a Bayesian optimization task.
arXiv Detail & Related papers (2022-02-01T17:46:51Z) - BCDAG: An R package for Bayesian structure and Causal learning of
Gaussian DAGs [77.34726150561087]
We introduce the R package for causal discovery and causal effect estimation from observational data.
Our implementation scales efficiently with the number of observations and, whenever the DAGs are sufficiently sparse, the number of variables in the dataset.
We then illustrate the main functions and algorithms on both real and simulated datasets.
arXiv Detail & Related papers (2022-01-28T09:30:32Z) - A Framework for an Assessment of the Kernel-target Alignment in Tree
Ensemble Kernel Learning [2.28438857884398]
We show that for continuous targets good performance of the tree-based kernel learning is associated with strong kernel-target alignment.
We also show that well performing tree ensemble based kernels are characterized by strong target aligned components.
arXiv Detail & Related papers (2021-08-19T15:37:17Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z) - Kernel learning approaches for summarising and combining posterior
similarity matrices [68.8204255655161]
We build upon the notion of the posterior similarity matrix (PSM) in order to suggest new approaches for summarising the output of MCMC algorithms for Bayesian clustering models.
A key contribution of our work is the observation that PSMs are positive semi-definite, and hence can be used to define probabilistically-motivated kernel matrices.
arXiv Detail & Related papers (2020-09-27T14:16:14Z) - TREX: Tree-Ensemble Representer-Point Explanations [13.109852233032395]
TREX is an explanation system that provides instance-attribution explanations for tree ensembles.
Since tree ensembles are non-differentiable, we define a kernel that captures the structure of the specific tree ensemble.
The weights in the kernel expansion of the surrogate model are used to define the global or local importance of each training example.
arXiv Detail & Related papers (2020-09-11T17:06:40Z) - Random Forest (RF) Kernel for Regression, Classification and Survival [1.8275108630751844]
We elucidate the performance and properties of the data driven RF kernels used by regularized linear models.
We show that for continuous and survival targets, the RF kernels are competitive to RF in higher dimensional scenarios.
We also provide the results from real life data sets for the regression, classification and survival to illustrate how these insights may be leveraged in practice.
arXiv Detail & Related papers (2020-08-31T20:21:27Z) - Graph Neural Networks with Composite Kernels [60.81504431653264]
We re-interpret node aggregation from the perspective of kernel weighting.
We present a framework to consider feature similarity in an aggregation scheme.
We propose feature aggregation as the composition of the original neighbor-based kernel and a learnable kernel to encode feature similarities in a feature space.
arXiv Detail & Related papers (2020-05-16T04:44:29Z) - Embedding Graph Auto-Encoder for Graph Clustering [90.8576971748142]
Graph auto-encoder (GAE) models are based on semi-supervised graph convolution networks (GCN)
We design a specific GAE-based model for graph clustering to be consistent with the theory, namely Embedding Graph Auto-Encoder (EGAE)
EGAE consists of one encoder and dual decoders.
arXiv Detail & Related papers (2020-02-20T09:53:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.