Tightness of prescriptive tree-based mixed-integer optimization
formulations
- URL: http://arxiv.org/abs/2302.14744v1
- Date: Tue, 28 Feb 2023 16:44:10 GMT
- Title: Tightness of prescriptive tree-based mixed-integer optimization
formulations
- Authors: Max Biggs and Georgia Perakis
- Abstract summary: We focus on modeling the relationship between an input feature vector and the predicted outcome of a trained decision tree.
We propose tighter mixed-integer optimization formulations than those previously introduced.
- Score: 2.538209532048867
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We focus on modeling the relationship between an input feature vector and the
predicted outcome of a trained decision tree using mixed-integer optimization.
This can be used in many practical applications where a decision tree or tree
ensemble is incorporated into an optimization problem to model the predicted
outcomes of a decision. We propose tighter mixed-integer optimization
formulations than those previously introduced. Existing formulations can be
shown to have linear relaxations that have fractional extreme points, even for
the simple case of modeling a single decision tree. A formulation we propose,
based on a projected union of polyhedra approach, is ideal for a single
decision tree. While the formulation is generally not ideal for tree ensembles
or if additional constraints are added, it generally has fewer extreme points,
leading to a faster time to solve, particularly if the formulation has
relatively few trees. However, previous work has shown that formulations based
on a binary representation of the feature vector perform well computationally
and hence are attractive for use in practical applications. We present multiple
approaches to tighten existing formulations with binary vectors, and show that
fractional extreme points are removed when there are multiple splits on the
same feature. At an extreme, we prove that this results in ideal formulations
for tree ensembles modeling a one-dimensional feature vector. Building on this
result, we also show via numerical simulations that these additional
constraints result in significantly tighter linear relaxations when the feature
vector is low dimensional. We also present instances where the time to solve to
optimality is significantly improved using these formulations.
Related papers
- Soft regression trees: a model variant and a decomposition training algorithm [0.24578723416255752]
We propose a new variant of soft multivariate regression trees (SRTs) where, for every input vector, the prediction is defined as a linear regression associated to a single leaf node.
SRTs exhibit the conditional computational property, i.e., each prediction depends on a small number of nodes.
Experiments on 15 wellknown datasets indicate that our SRTs and decomposition algorithm yield higher accuracy and robustness compared with traditional soft regression trees.
arXiv Detail & Related papers (2025-01-10T13:06:36Z) - Free Lunch in the Forest: Functionally-Identical Pruning of Boosted Tree Ensembles [45.962492329047215]
We introduce a method to prune a tree ensemble into a reduced version that is "functionally identical" to the original model.
We formalize the problem of functionally identical pruning on ensembles, introduce an exact optimization model, and provide a fast yet highly effective method to prune large ensembles.
arXiv Detail & Related papers (2024-08-28T23:15:46Z) - A Unified Approach to Extract Interpretable Rules from Tree Ensembles via Integer Programming [2.1408617023874443]
Tree ensembles are very popular machine learning models, known for their effectiveness in supervised classification and regression tasks.
Our work aims to extract an optimized list of rules from a trained tree ensemble, providing the user with a condensed, interpretable model that retains most of the predictive power of the full model.
Our extensive computational experiments offer statistically significant evidence that our method is competitive with other rule extraction methods in terms of predictive performance and fidelity towards the tree ensemble.
arXiv Detail & Related papers (2024-06-30T22:33:47Z) - bsnsing: A decision tree induction method based on recursive optimal
boolean rule composition [2.28438857884398]
This paper proposes a new mixed-integer programming (MIP) formulation to optimize split rule selection in the decision tree induction process.
It develops an efficient search solver that is able to solve practical instances faster than commercial solvers.
arXiv Detail & Related papers (2022-05-30T17:13:57Z) - Robust Optimal Classification Trees Against Adversarial Examples [5.254093731341154]
We propose a collection of methods to train decision trees that are optimally robust against user-specified attack models.
We show that the min-max optimization problem that arises in adversarial learning can be solved using a single minimization formulation.
We also present a method that determines the upper bound on adversarial accuracy for any model using bipartite matching.
arXiv Detail & Related papers (2021-09-08T18:10:49Z) - Partition-based formulations for mixed-integer optimization of trained
ReLU neural networks [66.88252321870085]
This paper introduces a class of mixed-integer formulations for trained ReLU neural networks.
At one extreme, one partition per input recovers the convex hull of a node, i.e., the tightest possible formulation for each node.
arXiv Detail & Related papers (2021-02-08T17:27:34Z) - Convex Polytope Trees [57.56078843831244]
convex polytope trees (CPT) are proposed to expand the family of decision trees by an interpretable generalization of their decision boundary.
We develop a greedy method to efficiently construct CPT and scalable end-to-end training algorithms for the tree parameters when the tree structure is given.
arXiv Detail & Related papers (2020-10-21T19:38:57Z) - MurTree: Optimal Classification Trees via Dynamic Programming and Search [61.817059565926336]
We present a novel algorithm for learning optimal classification trees based on dynamic programming and search.
Our approach uses only a fraction of the time required by the state-of-the-art and can handle datasets with tens of thousands of instances.
arXiv Detail & Related papers (2020-07-24T17:06:55Z) - Generalized and Scalable Optimal Sparse Decision Trees [56.35541305670828]
We present techniques that produce optimal decision trees over a variety of objectives.
We also introduce a scalable algorithm that produces provably optimal results in the presence of continuous variables.
arXiv Detail & Related papers (2020-06-15T19:00:11Z) - ENTMOOT: A Framework for Optimization over Ensemble Tree Models [57.98561336670884]
ENTMOOT is a framework for integrating tree models into larger optimization problems.
We show how ENTMOOT allows a simple integration of tree models into decision-making and black-box optimization.
arXiv Detail & Related papers (2020-03-10T14:34:07Z) - MINA: Convex Mixed-Integer Programming for Non-Rigid Shape Alignment [77.38594866794429]
convex mixed-integer programming formulation for non-rigid shape matching.
We propose a novel shape deformation model based on an efficient low-dimensional discrete model.
arXiv Detail & Related papers (2020-02-28T09:54:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.