GP-BART: a novel Bayesian additive regression trees approach using
Gaussian processes
- URL: http://arxiv.org/abs/2204.02112v4
- Date: Thu, 14 Sep 2023 18:27:18 GMT
- Title: GP-BART: a novel Bayesian additive regression trees approach using
Gaussian processes
- Authors: Mateus Maia, Keefe Murphy, Andrew C. Parnell
- Abstract summary: The GP-BART model is an extension of BART which addresses the limitation by assuming GP priors for the predictions of each terminal node among all trees.
The model's effectiveness is demonstrated through applications to simulated and real-world data, surpassing the performance of traditional modeling approaches in various scenarios.
- Score: 1.03590082373586
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Bayesian additive regression trees (BART) model is an ensemble method
extensively and successfully used in regression tasks due to its consistently
strong predictive performance and its ability to quantify uncertainty. BART
combines "weak" tree models through a set of shrinkage priors, whereby each
tree explains a small portion of the variability in the data. However, the lack
of smoothness and the absence of an explicit covariance structure over the
observations in standard BART can yield poor performance in cases where such
assumptions would be necessary. The Gaussian processes Bayesian additive
regression trees (GP-BART) model is an extension of BART which addresses this
limitation by assuming Gaussian process (GP) priors for the predictions of each
terminal node among all trees. The model's effectiveness is demonstrated
through applications to simulated and real-world data, surpassing the
performance of traditional modeling approaches in various scenarios.
Related papers
- Theory of Posterior Concentration for Generalized Bayesian Additive
Regression Trees [0.685316573653194]
We describe a Generalized framework for Bayesian trees and their additive ensembles.
We derive sufficient conditions on the response distribution, under which the posterior concentrates at a minimax rate, up to a logarithmic factor.
arXiv Detail & Related papers (2023-04-25T00:52:48Z) - Training Discrete Deep Generative Models via Gapped Straight-Through
Estimator [72.71398034617607]
We propose a Gapped Straight-Through ( GST) estimator to reduce the variance without incurring resampling overhead.
This estimator is inspired by the essential properties of Straight-Through Gumbel-Softmax.
Experiments demonstrate that the proposed GST estimator enjoys better performance compared to strong baselines on two discrete deep generative modeling tasks.
arXiv Detail & Related papers (2022-06-15T01:46:05Z) - Local Gaussian process extrapolation for BART models with applications
to causal inference [0.7734726150561088]
This paper proposes a novel extrapolation strategy that grafts Gaussian processes to the leaf nodes in BART for predicting points outside the range of the observed data.
In simulations studies, the new approach boasts superior performance compared to popular alternatives, such as Jackknife+.
arXiv Detail & Related papers (2022-04-23T00:37:53Z) - Hierarchical Embedded Bayesian Additive Regression Trees [0.0]
HE-BART allows for random effects to be included at the terminal node level of a set of regression trees.
Using simulated and real-world examples, we demonstrate that HE-BART yields superior predictions for many of the standard mixed effects models' example data sets.
In a future version of this paper, we outline its use in larger, more advanced data sets and structures.
arXiv Detail & Related papers (2022-04-14T19:56:03Z) - Distributional Gradient Boosting Machines [77.34726150561087]
Our framework is based on XGBoost and LightGBM.
We show that our framework achieves state-of-the-art forecast accuracy.
arXiv Detail & Related papers (2022-04-02T06:32:19Z) - Generalized Bayesian Additive Regression Trees Models: Beyond
Conditional Conjugacy [2.969705152497174]
In this article, we greatly expand the domain of applicability of BART to arbitrary emphgeneralized BART models.
Our algorithm requires only that the user be able to compute the likelihood and (optionally) its gradient and Fisher information.
The potential applications are very broad; we consider examples in survival analysis, structured heteroskedastic regression, and gamma shape regression.
arXiv Detail & Related papers (2022-02-20T22:52:07Z) - BCD Nets: Scalable Variational Approaches for Bayesian Causal Discovery [97.79015388276483]
A structural equation model (SEM) is an effective framework to reason over causal relationships represented via a directed acyclic graph (DAG)
Recent advances enabled effective maximum-likelihood point estimation of DAGs from observational data.
We propose BCD Nets, a variational framework for estimating a distribution over DAGs characterizing a linear-Gaussian SEM.
arXiv Detail & Related papers (2021-12-06T03:35:21Z) - A cautionary tale on fitting decision trees to data from additive
models: generalization lower bounds [9.546094657606178]
We study the generalization performance of decision trees with respect to different generative regression models.
This allows us to elicit their inductive bias, that is, the assumptions the algorithms make (or do not make) to generalize to new data.
We prove a sharp squared error generalization lower bound for a large class of decision tree algorithms fitted to sparse additive models.
arXiv Detail & Related papers (2021-10-18T21:22:40Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z) - Likelihood-Free Inference with Deep Gaussian Processes [70.74203794847344]
Surrogate models have been successfully used in likelihood-free inference to decrease the number of simulator evaluations.
We propose a Deep Gaussian Process (DGP) surrogate model that can handle more irregularly behaved target distributions.
Our experiments show how DGPs can outperform GPs on objective functions with multimodal distributions and maintain a comparable performance in unimodal cases.
arXiv Detail & Related papers (2020-06-18T14:24:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.