Theory of Posterior Concentration for Generalized Bayesian Additive
Regression Trees
- URL: http://arxiv.org/abs/2304.12505v1
- Date: Tue, 25 Apr 2023 00:52:48 GMT
- Title: Theory of Posterior Concentration for Generalized Bayesian Additive
Regression Trees
- Authors: Enakshi Saha
- Abstract summary: We describe a Generalized framework for Bayesian trees and their additive ensembles.
We derive sufficient conditions on the response distribution, under which the posterior concentrates at a minimax rate, up to a logarithmic factor.
- Score: 0.685316573653194
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Bayesian Additive Regression Trees (BART) are a powerful semiparametric
ensemble learning technique for modeling nonlinear regression functions.
Although initially BART was proposed for predicting only continuous and binary
response variables, over the years multiple extensions have emerged that are
suitable for estimating a wider class of response variables (e.g. categorical
and count data) in a multitude of application areas. In this paper we describe
a Generalized framework for Bayesian trees and their additive ensembles where
the response variable comes from an exponential family distribution and hence
encompasses a majority of these variants of BART. We derive sufficient
conditions on the response distribution, under which the posterior concentrates
at a minimax rate, up to a logarithmic factor. In this regard our results
provide theoretical justification for the empirical success of BART and its
variants.
Related papers
- A Non-negative VAE:the Generalized Gamma Belief Network [49.970917207211556]
The gamma belief network (GBN) has demonstrated its potential for uncovering multi-layer interpretable latent representations in text data.
We introduce the generalized gamma belief network (Generalized GBN) in this paper, which extends the original linear generative model to a more expressive non-linear generative model.
We also propose an upward-downward Weibull inference network to approximate the posterior distribution of the latent variables.
arXiv Detail & Related papers (2024-08-06T18:18:37Z) - The Computational Curse of Big Data for Bayesian Additive Regression Trees: A Hitting Time Analysis [8.36826153664925]
We show that the BART sampler often converges slowly, confirming empirical observations by other researchers.
As $n$ increases, the approximate BART posterior becomes increasingly different from the exact posterior.
arXiv Detail & Related papers (2024-06-28T14:45:29Z) - Conformal inference for regression on Riemannian Manifolds [49.7719149179179]
We investigate prediction sets for regression scenarios when the response variable, denoted by $Y$, resides in a manifold, and the covariable, denoted by X, lies in Euclidean space.
We prove the almost sure convergence of the empirical version of these regions on the manifold to their population counterparts.
arXiv Detail & Related papers (2023-10-12T10:56:25Z) - flexBART: Flexible Bayesian regression trees with categorical predictors [0.6577148087211809]
Most implementations of Bayesian additive regression trees (BART) one-hot encode categorical predictors, replacing each one with several binary indicators.
We re-implement BART with regression trees that can assign multiple levels to both branches of a decision tree node.
Our re-implementation, which is available in the flexBART package, often yields improved out-of-sample predictive performance and scales better to larger datasets.
arXiv Detail & Related papers (2022-11-08T18:52:37Z) - Instance-Dependent Generalization Bounds via Optimal Transport [51.71650746285469]
Existing generalization bounds fail to explain crucial factors that drive the generalization of modern neural networks.
We derive instance-dependent generalization bounds that depend on the local Lipschitz regularity of the learned prediction function in the data space.
We empirically analyze our generalization bounds for neural networks, showing that the bound values are meaningful and capture the effect of popular regularization methods during training.
arXiv Detail & Related papers (2022-11-02T16:39:42Z) - Training Discrete Deep Generative Models via Gapped Straight-Through
Estimator [72.71398034617607]
We propose a Gapped Straight-Through ( GST) estimator to reduce the variance without incurring resampling overhead.
This estimator is inspired by the essential properties of Straight-Through Gumbel-Softmax.
Experiments demonstrate that the proposed GST estimator enjoys better performance compared to strong baselines on two discrete deep generative modeling tasks.
arXiv Detail & Related papers (2022-06-15T01:46:05Z) - GP-BART: a novel Bayesian additive regression trees approach using
Gaussian processes [1.03590082373586]
The GP-BART model is an extension of BART which addresses the limitation by assuming GP priors for the predictions of each terminal node among all trees.
The model's effectiveness is demonstrated through applications to simulated and real-world data, surpassing the performance of traditional modeling approaches in various scenarios.
arXiv Detail & Related papers (2022-04-05T11:18:44Z) - An Intermediate-level Attack Framework on The Basis of Linear Regression [89.85593878754571]
This paper substantially extends our work published at ECCV, in which an intermediate-level attack was proposed to improve the transferability of some baseline adversarial examples.
We advocate to establish a direct linear mapping from the intermediate-level discrepancies (between adversarial features and benign features) to classification prediction loss of the adversarial example.
We show that 1) a variety of linear regression models can all be considered in order to establish the mapping, 2) the magnitude of the finally obtained intermediate-level discrepancy is linearly correlated with adversarial transferability, and 3) further boost of the performance can be achieved by performing multiple runs of the baseline attack with
arXiv Detail & Related papers (2022-03-21T03:54:53Z) - Generalized Bayesian Additive Regression Trees Models: Beyond
Conditional Conjugacy [2.969705152497174]
In this article, we greatly expand the domain of applicability of BART to arbitrary emphgeneralized BART models.
Our algorithm requires only that the user be able to compute the likelihood and (optionally) its gradient and Fisher information.
The potential applications are very broad; we consider examples in survival analysis, structured heteroskedastic regression, and gamma shape regression.
arXiv Detail & Related papers (2022-02-20T22:52:07Z) - A Unified Framework for Multi-distribution Density Ratio Estimation [101.67420298343512]
Binary density ratio estimation (DRE) provides the foundation for many state-of-the-art machine learning algorithms.
We develop a general framework from the perspective of Bregman minimization divergence.
We show that our framework leads to methods that strictly generalize their counterparts in binary DRE.
arXiv Detail & Related papers (2021-12-07T01:23:20Z) - Accounting for shared covariates in semi-parametric Bayesian additive regression trees [0.0]
We propose some extensions to semi-parametric models based on Bayesian additive regression trees (BART)
The main novelty in our approach lies in the way we change the tree-generation moves in BART to deal with this bias.
We show competitive performance when compared to regression models, alternative formulations of semi-parametric BART, and other tree-based methods.
arXiv Detail & Related papers (2021-08-17T13:58:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.