Learning to Branch in Combinatorial Optimization with Graph Pointer
Networks
- URL: http://arxiv.org/abs/2307.01434v1
- Date: Tue, 4 Jul 2023 01:56:07 GMT
- Title: Learning to Branch in Combinatorial Optimization with Graph Pointer
Networks
- Authors: Rui Wang, Zhiming Zhou, Tao Zhang, Ling Wang, Xin Xu, Xiangke Liao,
Kaiwen Li
- Abstract summary: This paper proposes a graph pointer network model for learning the variable selection policy in the branch-and-bound.
The proposed model, which combines the graph neural network and the pointer mechanism, can effectively map from the solver state to the branching variable decisions.
- Score: 17.729352126574902
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Branch-and-bound is a typical way to solve combinatorial optimization
problems. This paper proposes a graph pointer network model for learning the
variable selection policy in the branch-and-bound. We extract the graph
features, global features and historical features to represent the solver
state. The proposed model, which combines the graph neural network and the
pointer mechanism, can effectively map from the solver state to the branching
variable decisions. The model is trained to imitate the classic strong
branching expert rule by a designed top-k Kullback-Leibler divergence loss
function. Experiments on a series of benchmark problems demonstrate that the
proposed approach significantly outperforms the widely used expert-designed
branching rules. Our approach also outperforms the state-of-the-art
machine-learning-based branch-and-bound methods in terms of solving speed and
search tree size on all the test instances. In addition, the model can
generalize to unseen instances and scale to larger instances.
Related papers
- A Benchmark for Maximum Cut: Towards Standardization of the Evaluation of Learned Heuristics for Combinatorial Optimization [12.016449555335976]
We propose an open-source benchmark suite MaxCut-Bench dedicated to the NP-hard Maximum Cut problem in both its weighted and unweighted variants.
We use the benchmark in an attempt to systematically corroborate or reproduce the results of several, popular learning-based approaches.
Our results show that several of the learneds fail to outperform a naive greedy algorithm, and that only one of them consistently outperforms Tabu Search.
arXiv Detail & Related papers (2024-06-14T19:44:23Z) - Graph Convolutional Branch and Bound [1.8966938152549224]
This article demonstrates the effectiveness of employing a deep learning model in an optimization pipeline.
In this context, neural networks can be leveraged to rapidly acquire valuable information.
arXiv Detail & Related papers (2024-06-05T09:42:43Z) - Ensemble Quadratic Assignment Network for Graph Matching [52.20001802006391]
Graph matching is a commonly used technique in computer vision and pattern recognition.
Recent data-driven approaches have improved the graph matching accuracy remarkably.
We propose a graph neural network (GNN) based approach to combine the advantages of data-driven and traditional methods.
arXiv Detail & Related papers (2024-03-11T06:34:05Z) - Optimizing Neural Networks with Gradient Lexicase Selection [7.554281369517276]
A potential drawback of using aggregated performance measurement in machine learning is that models may learn to accept higher errors on some training cases as compromises for lower errors on others.
We propose Gradient Lexicase Selection, an optimization framework that combines gradient descent and lexicase selection in an evolutionary fashion.
Our experimental results demonstrate that the proposed method improves the generalization performance of various widely-used deep neural network architectures.
arXiv Detail & Related papers (2023-12-19T21:21:25Z) - Discrete Graph Auto-Encoder [52.50288418639075]
We introduce a new framework named Discrete Graph Auto-Encoder (DGAE)
We first use a permutation-equivariant auto-encoder to convert graphs into sets of discrete latent node representations.
In the second step, we sort the sets of discrete latent representations and learn their distribution with a specifically designed auto-regressive model.
arXiv Detail & Related papers (2023-06-13T12:40:39Z) - Learning Branching Heuristics from Graph Neural Networks [1.4660170768702356]
We first propose a new graph neural network (GNN) model designed using a probabilistic method.
Our approach introduces a new way of applying GNNs towards enhancing the classical backtracking algorithm used in AI.
arXiv Detail & Related papers (2022-11-26T00:01:01Z) - Equivariance with Learned Canonicalization Functions [77.32483958400282]
We show that learning a small neural network to perform canonicalization is better than using predefineds.
Our experiments show that learning the canonicalization function is competitive with existing techniques for learning equivariant functions across many tasks.
arXiv Detail & Related papers (2022-11-11T21:58:15Z) - Learning to Compare Nodes in Branch and Bound with Graph Neural Networks [5.08128537391027]
Branch-and-bound approaches in integer programming require ordering portions of the space to explore next.
We propose a new siamese graph neural network model to tackle this problem, where the nodes are represented as bipartite graphs with attributes.
We evaluate our method by solving the instances in a plain framework where the nodes are explored according to their rank.
arXiv Detail & Related papers (2022-10-30T19:38:23Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Temporal Graph Network Embedding with Causal Anonymous Walks
Representations [54.05212871508062]
We propose a novel approach for dynamic network representation learning based on Temporal Graph Network.
For evaluation, we provide a benchmark pipeline for the evaluation of temporal network embeddings.
We show the applicability and superior performance of our model in the real-world downstream graph machine learning task provided by one of the top European banks.
arXiv Detail & Related papers (2021-08-19T15:39:52Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.