Learning to branch with Tree MDPs
- URL: http://arxiv.org/abs/2205.11107v1
- Date: Mon, 23 May 2022 07:57:32 GMT
- Title: Learning to branch with Tree MDPs
- Authors: Lara Scavuzzo and Feng Yang Chen and Didier Ch\'etelat and Maxime
Gasse and Andrea Lodi and Neil Yorke-Smith and Karen Aardal
- Abstract summary: We propose to learn branching rules from scratch via Reinforcement Learning (RL)
We propose tree Markov Decision Processes, or tree MDPs, a generalization of temporal MDPs that provides a more suitable framework for learning to branch.
We demonstrate through computational experiments that tree MDPs improve the learning convergence, and offer a promising framework for tackling the learning-to-branch problem in MILPs.
- Score: 6.754135838894833
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: State-of-the-art Mixed Integer Linear Program (MILP) solvers combine
systematic tree search with a plethora of hard-coded heuristics, such as the
branching rule. The idea of learning branching rules from data has received
increasing attention recently, and promising results have been obtained by
learning fast approximations of the strong branching expert. In this work, we
instead propose to learn branching rules from scratch via Reinforcement
Learning (RL). We revisit the work of Etheve et al. (2020) and propose tree
Markov Decision Processes, or tree MDPs, a generalization of temporal MDPs that
provides a more suitable framework for learning to branch. We derive a tree
policy gradient theorem, which exhibits a better credit assignment compared to
its temporal counterpart. We demonstrate through computational experiments that
tree MDPs improve the learning convergence, and offer a promising framework for
tackling the learning-to-branch problem in MILPs.
Related papers
- TreeDQN: Learning to minimize Branch-and-Bound tree [78.52895577861327]
Branch-and-Bound is a convenient approach to solving optimization tasks in the form of Mixed Linear Programs.
The efficiency of the solver depends on the branchning used to select a variable for splitting.
We propose a reinforcement learning method that can efficiently learn the branching.
arXiv Detail & Related papers (2023-06-09T14:01:26Z) - RLET: A Reinforcement Learning Based Approach for Explainable QA with
Entailment Trees [47.745218107037786]
We propose RLET, a Reinforcement Learning based Entailment Tree generation framework.
RLET iteratively performs single step reasoning with sentence selection and deduction generation modules.
Experiments on three settings of the EntailmentBank dataset demonstrate the strength of using RL framework.
arXiv Detail & Related papers (2022-10-31T06:45:05Z) - Branch Ranking for Efficient Mixed-Integer Programming via Offline
Ranking-based Policy Learning [45.1011106869493]
We formulate learning to branch as an offline reinforcement learning (RL) problem.
We train a branching model named Branch Ranking via offline policy learning.
Experiments on synthetic MIP benchmarks and real-world tasks demonstrate that Branch Rankink is more efficient and robust.
arXiv Detail & Related papers (2022-07-26T15:34:10Z) - Reinforcement Learning for Branch-and-Bound Optimisation using
Retrospective Trajectories [72.15369769265398]
Machine learning has emerged as a promising paradigm for branching.
We propose retro branching; a simple yet effective approach to RL for branching.
We outperform the current state-of-the-art RL branching algorithm by 3-5x and come within 20% of the best IL method's performance on MILPs with 500 constraints and 1000 variables.
arXiv Detail & Related papers (2022-05-28T06:08:07Z) - A Closer Look at Branch Classifiers of Multi-exit Architectures [103.27533521196817]
Constant-complexity branching keeps all branches the same, while complexity-increasing and complexity-decreasing branching place more complex branches later or earlier in the backbone respectively.
We investigate a cause by using knowledge consistency to probe the effect of adding branches onto a backbone.
Our findings show that complexity-decreasing branching yields the least disruption to the feature abstraction hierarchy of the backbone.
arXiv Detail & Related papers (2022-04-28T08:37:25Z) - MurTree: Optimal Classification Trees via Dynamic Programming and Search [61.817059565926336]
We present a novel algorithm for learning optimal classification trees based on dynamic programming and search.
Our approach uses only a fraction of the time required by the state-of-the-art and can handle datasets with tens of thousands of instances.
arXiv Detail & Related papers (2020-07-24T17:06:55Z) - Parameterizing Branch-and-Bound Search Trees to Learn Branching Policies [76.83991682238666]
Branch and Bound (B&B) is the exact tree search method typically used to solve Mixed-Integer Linear Programming problems (MILPs)
We propose a novel imitation learning framework, and introduce new input features and architectures to represent branching.
arXiv Detail & Related papers (2020-02-12T17:43:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.