The Neural-Prediction based Acceleration Algorithm of Column Generation
for Graph-Based Set Covering Problems
- URL: http://arxiv.org/abs/2207.01411v1
- Date: Mon, 4 Jul 2022 13:46:59 GMT
- Title: The Neural-Prediction based Acceleration Algorithm of Column Generation
for Graph-Based Set Covering Problems
- Authors: Haofeng Yuan, Peng Jiang and Shiji Song
- Abstract summary: We propose an improved column generation algorithm with neural prediction (CG-P) for solving graph-based set covering problems.
We leverage a graph neural network based neural prediction model to predict the probability to be included in the final solution for each edge.
We evaluate the CG-P algorithm on railway crew scheduling problems and it outperforms the baseline column generation algorithm.
- Score: 20.1479227701035
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Set covering problem is an important class of combinatorial optimization
problems, which has been widely applied and studied in many fields. In this
paper, we propose an improved column generation algorithm with neural
prediction (CG-P) for solving graph-based set covering problems. We leverage a
graph neural network based neural prediction model to predict the probability
to be included in the final solution for each edge. Our CG-P algorithm
constructs a reduced graph that only contains the edges with higher predicted
probability, and this graph reduction process significantly speeds up the
solution process. We evaluate the CG-P algorithm on railway crew scheduling
problems and it outperforms the baseline column generation algorithm. We
provide two solution modes for our CG-P algorithm. In the optimal mode, we can
obtain a solution with an optimality guarantee while reducing the time cost to
63.12%. In the fast mode, we can obtain a sub-optimal solution with a 7.62%
optimality gap in only 2.91% computation time.
Related papers
- Decision-focused Graph Neural Networks for Combinatorial Optimization [62.34623670845006]
An emerging strategy to tackle optimization problems involves the adoption of graph neural networks (GNNs) as an alternative to traditional algorithms.
Despite the growing popularity of GNNs and traditional algorithm solvers in the realm of CO, there is limited research on their integrated use and the correlation between them within an end-to-end framework.
We introduce a decision-focused framework that utilizes GNNs to address CO problems with auxiliary support.
arXiv Detail & Related papers (2024-06-05T22:52:27Z) - Are Graph Neural Networks Optimal Approximation Algorithms? [26.5364112420121]
We design graph neural network architectures that capture optimal approximation algorithms for a class of optimization problems.
We take advantage of OptGNN's ability to capture convex relaxations to design an algorithm for producing bounds on the optimal solution from the learned embeddings of OptGNN.
arXiv Detail & Related papers (2023-10-01T00:12:31Z) - Unfolded proximal neural networks for robust image Gaussian denoising [7.018591019975253]
We propose a unified framework to build PNNs for the Gaussian denoising task, based on both the dual-FB and the primal-dual Chambolle-Pock algorithms.
We also show that accelerated versions of these algorithms enable skip connections in the associated NN layers.
arXiv Detail & Related papers (2023-08-06T15:32:16Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - CBAG: An Efficient Genetic Algorithm for the Graph Burning Problem [0.0]
We propose an efficient genetic algorithm called Centrality BAsed Genetic-algorithm (CBAG) for solving the graph burning problem.
Considering the unique characteristics of the graph burning problem, we introduce novel chromosome representation, and evaluation method.
Based on the results, it can be seen that the proposed algorithm achieves better performance in comparison to the previous state-of-the-art centralitys.
arXiv Detail & Related papers (2022-08-01T17:34:07Z) - Neural Improvement Heuristics for Graph Combinatorial Optimization
Problems [49.85111302670361]
We introduce a novel Neural Improvement (NI) model capable of handling graph-based problems where information is encoded in the nodes, edges, or both.
The presented model serves as a fundamental component for hill-climbing-based algorithms that guide the selection of neighborhood operations for each.
arXiv Detail & Related papers (2022-06-01T10:35:29Z) - Improved Acyclicity Reasoning for Bayesian Network Structure Learning
with Constraint Programming [0.0]
Learning the structure of a Bayesian network (BNSL) from discrete data is known to be an NP-hard task.
In this work, we propose a new time algorithm for discovering a subset of all possible cluster cuts.
We show that, despite being suboptimal, they improve performance by orders of magnitude.
The resulting solver also compares favourably with GOBNILP, a state-of-the-art solver for the BNSL problem.
arXiv Detail & Related papers (2021-06-23T09:46:11Z) - A Bi-Level Framework for Learning to Solve Combinatorial Optimization on
Graphs [91.07247251502564]
We propose a hybrid approach to combine the best of the two worlds, in which a bi-level framework is developed with an upper-level learning method to optimize the graph.
Such a bi-level approach simplifies the learning on the original hard CO and can effectively mitigate the demand for model capacity.
arXiv Detail & Related papers (2021-06-09T09:18:18Z) - An Asymptotically Optimal Primal-Dual Incremental Algorithm for
Contextual Linear Bandits [129.1029690825929]
We introduce a novel algorithm improving over the state-of-the-art along multiple dimensions.
We establish minimax optimality for any learning horizon in the special case of non-contextual linear bandits.
arXiv Detail & Related papers (2020-10-23T09:12:47Z) - Private Stochastic Convex Optimization: Optimal Rates in Linear Time [74.47681868973598]
We study the problem of minimizing the population loss given i.i.d. samples from a distribution over convex loss functions.
A recent work of Bassily et al. has established the optimal bound on the excess population loss achievable given $n$ samples.
We describe two new techniques for deriving convex optimization algorithms both achieving the optimal bound on excess loss and using $O(minn, n2/d)$ gradient computations.
arXiv Detail & Related papers (2020-05-10T19:52:03Z) - Gumbel-softmax-based Optimization: A Simple General Framework for
Optimization Problems on Graphs [5.486093983007419]
We propose a simple, fast, and general algorithm framework based on advanced automatic differentiation technique empowered by deep learning frameworks.
High-quality solutions can be obtained with much less time consuming compared to traditional approaches.
arXiv Detail & Related papers (2020-04-14T14:11:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.