RJHMC-Tree for Exploration of the Bayesian Decision Tree Posterior
- URL: http://arxiv.org/abs/2312.01577v1
- Date: Mon, 4 Dec 2023 02:23:32 GMT
- Title: RJHMC-Tree for Exploration of the Bayesian Decision Tree Posterior
- Authors: Jodie A. Cochrane, Adrian G. Wills, Sarah J. Johnson
- Abstract summary: This paper is directed towards learning decision trees from data using a Bayesian approach.
It investigates using a Hamiltonian Monte Carlo (HMC) approach to explore the posterior of Bayesian decision trees more efficiently.
- Score: 1.3351610617039973
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Decision trees have found widespread application within the machine learning
community due to their flexibility and interpretability. This paper is directed
towards learning decision trees from data using a Bayesian approach, which is
challenging due to the potentially enormous parameter space required to span
all tree models. Several approaches have been proposed to combat this
challenge, with one of the more successful being Markov chain Monte Carlo
(MCMC) methods. The efficacy and efficiency of MCMC methods fundamentally rely
on the quality of the so-called proposals, which is the focus of this paper. In
particular, this paper investigates using a Hamiltonian Monte Carlo (HMC)
approach to explore the posterior of Bayesian decision trees more efficiently
by exploiting the geometry of the likelihood within a global update scheme. Two
implementations of the novel algorithm are developed and compared to existing
methods by testing against standard datasets in the machine learning and
Bayesian decision tree literature. HMC-based methods are shown to perform
favourably with respect to predictive test accuracy, acceptance rate, and tree
complexity.
Related papers
- Technical Report: Enhancing LLM Reasoning with Reward-guided Tree Search [95.06503095273395]
o1-like reasoning approach is challenging, and researchers have been making various attempts to advance this open area of research.
We present a preliminary exploration into enhancing the reasoning abilities of LLMs through reward-guided tree search algorithms.
arXiv Detail & Related papers (2024-11-18T16:15:17Z) - LLaMA-Berry: Pairwise Optimization for O1-like Olympiad-Level Mathematical Reasoning [56.273799410256075]
The framework combines Monte Carlo Tree Search (MCTS) with iterative Self-Refine to optimize the reasoning path.
The framework has been tested on general and advanced benchmarks, showing superior performance in terms of search efficiency and problem-solving capability.
arXiv Detail & Related papers (2024-10-03T18:12:29Z) - Learning Deep Tree-based Retriever for Efficient Recommendation: Theory and Method [76.31185707649227]
We propose a Deep Tree-based Retriever (DTR) for efficient recommendation.
DTR frames the training task as a softmax-based multi-class classification over tree nodes at the same level.
To mitigate the suboptimality induced by the labeling of non-leaf nodes, we propose a rectification method for the loss function.
arXiv Detail & Related papers (2024-08-21T05:09:53Z) - Learning accurate and interpretable decision trees [27.203303726977616]
We develop approaches to design decision tree learning algorithms given repeated access to data from the same domain.
We study the sample complexity of tuning prior parameters in Bayesian decision tree learning, and extend our results to decision tree regression.
We also study the interpretability of the learned decision trees and introduce a data-driven approach for optimizing the explainability versus accuracy trade-off using decision trees.
arXiv Detail & Related papers (2024-05-24T20:10:10Z) - Variational Bayesian Methods for a Tree-Structured Stick-Breaking Process Mixture of Gaussians by Application of the Bayes Codes for Context Tree Models [0.0]
We propose a learning algorithm with less computational cost for the TS-SBP mixture of Gaussians.
The main challenge is efficient calculation of a sum over all possible trees.
arXiv Detail & Related papers (2024-05-01T08:36:13Z) - Divide, Conquer, Combine Bayesian Decision Tree Sampling [1.1879716317856945]
Decision trees are commonly used predictive models due to their flexibility and interpretability.
This paper is directed at quantifying the uncertainty of decision tree predictions by employing a Bayesian inference approach.
arXiv Detail & Related papers (2024-03-26T23:14:15Z) - Bayesian Decision Trees Inspired from Evolutionary Algorithms [64.80360020499555]
We propose a replacement of the Markov Chain Monte Carlo (MCMC) with an inherently parallel algorithm, the Sequential Monte Carlo (SMC)
Experiments show that SMC combined with the Evolutionary Algorithms (EA) can produce more accurate results compared to MCMC in 100 times fewer iterations.
arXiv Detail & Related papers (2023-05-30T06:17:35Z) - Bayesian Decision Trees via Tractable Priors and Probabilistic
Context-Free Grammars [7.259767735431625]
We propose a new criterion for training Bayesian Decision Trees.
BCART-PCFG can efficiently sample decision trees from a posterior distribution across trees given the data.
We find that trees sampled via BCART-PCFG perform comparable to or better than greedily-constructed Decision Trees.
arXiv Detail & Related papers (2023-02-15T00:17:41Z) - An Efficient Dynamic Sampling Policy For Monte Carlo Tree Search [0.0]
We consider the popular tree-based search strategy within the framework of reinforcement learning, the Monte Carlo Tree Search (MCTS)
We propose a dynamic sampling tree policy that efficiently allocates limited computational budget to maximize the probability of correct selection of the best action at the root node of the tree.
arXiv Detail & Related papers (2022-04-26T02:39:18Z) - Rectified Decision Trees: Exploring the Landscape of Interpretable and
Effective Machine Learning [66.01622034708319]
We propose a knowledge distillation based decision trees extension, dubbed rectified decision trees (ReDT)
We extend the splitting criteria and the ending condition of the standard decision trees, which allows training with soft labels.
We then train the ReDT based on the soft label distilled from a well-trained teacher model through a novel jackknife-based method.
arXiv Detail & Related papers (2020-08-21T10:45:25Z) - MurTree: Optimal Classification Trees via Dynamic Programming and Search [61.817059565926336]
We present a novel algorithm for learning optimal classification trees based on dynamic programming and search.
Our approach uses only a fraction of the time required by the state-of-the-art and can handle datasets with tens of thousands of instances.
arXiv Detail & Related papers (2020-07-24T17:06:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.