Optimal Policy Trees
- URL: http://arxiv.org/abs/2012.02279v1
- Date: Thu, 3 Dec 2020 21:41:22 GMT
- Title: Optimal Policy Trees
- Authors: Maxime Amram, Jack Dunn, Ying Daisy Zhuo
- Abstract summary: We propose an approach for learning optimal tree-based prescription policies directly from data.
We combine methods for counterfactual estimation from the causal inference literature with recent advances in training globally-optimal decision trees.
The resulting method, Optimal Policy Trees, yields interpretable prescription policies.
- Score: 0.6015898117103068
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose an approach for learning optimal tree-based prescription policies
directly from data, combining methods for counterfactual estimation from the
causal inference literature with recent advances in training globally-optimal
decision trees. The resulting method, Optimal Policy Trees, yields
interpretable prescription policies, is highly scalable, and handles both
discrete and continuous treatments. We conduct extensive experiments on both
synthetic and real-world datasets and demonstrate that these trees offer
best-in-class performance across a wide variety of problems.
Related papers
- Learning a Decision Tree Algorithm with Transformers [75.96920867382859]
We introduce MetaTree, a transformer-based model trained via meta-learning to directly produce strong decision trees.
We fit both greedy decision trees and globally optimized decision trees on a large number of datasets, and train MetaTree to produce only the trees that achieve strong generalization performance.
arXiv Detail & Related papers (2024-02-06T07:40:53Z) - Rolling Lookahead Learning for Optimal Classification Trees [0.0]
We propose a rolling subtree lookahead algorithm that combines the relative scalability of the myopic approaches with the foresight of the optimal approaches in constructing trees.
We show that our approach outperforms optimal and myopic approaches in 808 out of 1330 problem instances, improving the out-of-sample accuracy by up to 23.6% and 14.4%, respectively.
arXiv Detail & Related papers (2023-04-21T09:17:00Z) - A Mathematical Programming Approach to Optimal Classification Forests [1.0705399532413618]
We propose a novel mathematical optimization-based methodology in which a given number of trees are simultaneously constructed.
The classification rule is derived by assigning to each observation its most frequently predicted class among the trees in the forest.
We show that our proposed method has equal or superior performance compared with state-of-the-art tree-based classification methods.
arXiv Detail & Related papers (2022-11-18T20:33:08Z) - Social Interpretable Tree for Pedestrian Trajectory Prediction [75.81745697967608]
We propose a tree-based method, termed as Social Interpretable Tree (SIT), to address this multi-modal prediction task.
A path in the tree from the root to leaf represents an individual possible future trajectory.
Despite the hand-crafted tree, the experimental results on ETH-UCY and Stanford Drone datasets demonstrate that our method is capable of matching or exceeding the performance of state-of-the-art methods.
arXiv Detail & Related papers (2022-05-26T12:18:44Z) - POETREE: Interpretable Policy Learning with Adaptive Decision Trees [78.6363825307044]
POETREE is a novel framework for interpretable policy learning.
It builds probabilistic tree policies determining physician actions based on patients' observations and medical history.
It outperforms the state-of-the-art on real and synthetic medical datasets.
arXiv Detail & Related papers (2022-03-15T16:50:52Z) - Learning Optimal Prescriptive Trees from Observational Data [7.215903549622416]
We propose a method for learning optimal prescriptive trees using mixed-integer optimization (MIO) technology.
Contrary to existing literature, our approach does not require data to be randomized, 2) does not impose stringent assumptions on the learned trees, and 3) has the ability to model domain specific constraints.
arXiv Detail & Related papers (2021-08-31T05:38:36Z) - Strong Optimal Classification Trees [8.10995244893652]
We propose an intuitive flow-based MIO formulation for learning optimal binary classification trees.
Our formulation can accommodate side constraints to enable the design of interpretable and fair decision trees.
We show that our proposed approaches are 29 times faster than state-of-the-art MIO-based techniques.
arXiv Detail & Related papers (2021-03-29T21:40:58Z) - Stochastic Optimization Forests [60.523606291705214]
We show how to train forest decision policies by growing trees that choose splits to directly optimize the downstream decision quality, rather than splitting to improve prediction accuracy as in the standard random forest algorithm.
We show that our approximate splitting criteria can reduce running time hundredfold, while achieving performance close to forest algorithms that exactly re-optimize for every candidate split.
arXiv Detail & Related papers (2020-08-17T16:56:06Z) - MurTree: Optimal Classification Trees via Dynamic Programming and Search [61.817059565926336]
We present a novel algorithm for learning optimal classification trees based on dynamic programming and search.
Our approach uses only a fraction of the time required by the state-of-the-art and can handle datasets with tens of thousands of instances.
arXiv Detail & Related papers (2020-07-24T17:06:55Z) - Generalized and Scalable Optimal Sparse Decision Trees [56.35541305670828]
We present techniques that produce optimal decision trees over a variety of objectives.
We also introduce a scalable algorithm that produces provably optimal results in the presence of continuous variables.
arXiv Detail & Related papers (2020-06-15T19:00:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.