SeaPearl: A Constraint Programming Solver guided by Reinforcement
Learning
- URL: http://arxiv.org/abs/2102.09193v1
- Date: Thu, 18 Feb 2021 07:34:38 GMT
- Title: SeaPearl: A Constraint Programming Solver guided by Reinforcement
Learning
- Authors: F\'elix Chalumeau (1), Ilan Coulon (1), Quentin Cappart (2),
Louis-Martin Rousseau (2) ((1) \'Ecole Polytechnique, Institut Polytechnique
de Paris, (2) \'Ecole Polytechnique de Montr\'eal)
- Abstract summary: This paper presents the proof of concept for SeaPearl, a new constraint programming solver implemented in Julia.
SeaPearl supports machine learning routines in order to learn branching decisions using reinforcement learning.
Although not yet competitive with industrial solvers, SeaPearl aims to provide a flexible and open-source framework.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The design of efficient and generic algorithms for solving combinatorial
optimization problems has been an active field of research for many years.
Standard exact solving approaches are based on a clever and complete
enumeration of the solution set. A critical and non-trivial design choice with
such methods is the branching strategy, directing how the search is performed.
The last decade has shown an increasing interest in the design of machine
learning-based heuristics to solve combinatorial optimization problems. The
goal is to leverage knowledge from historical data to solve similar new
instances of a problem. Used alone, such heuristics are only able to provide
approximate solutions efficiently, but cannot prove optimality nor bounds on
their solution. Recent works have shown that reinforcement learning can be
successfully used for driving the search phase of constraint programming (CP)
solvers. However, it has also been shown that this hybridization is challenging
to build, as standard CP frameworks do not natively include machine learning
mechanisms, leading to some sources of inefficiencies. This paper presents the
proof of concept for SeaPearl, a new CP solver implemented in Julia, that
supports machine learning routines in order to learn branching decisions using
reinforcement learning. Support for modeling the learning component is also
provided. We illustrate the modeling and solution performance of this new
solver on two problems. Although not yet competitive with industrial solvers,
SeaPearl aims to provide a flexible and open-source framework in order to
facilitate future research in the hybridization of constraint programming and
machine learning.
Related papers
- An Unsupervised Learning Framework Combined with Heuristics for the Maximum Minimal Cut Problem [5.092968949752308]
This work proposes an unsupervised learning framework combined with Maximums for MMCP.
A crucial observation is that each solution corresponds to at least one spanning tree.
We conduct extensive experiments to evaluate our framework and give a specific application.
arXiv Detail & Related papers (2024-08-16T02:07:34Z) - Machine Learning Insides OptVerse AI Solver: Design Principles and
Applications [74.67495900436728]
We present a comprehensive study on the integration of machine learning (ML) techniques into Huawei Cloud's OptVerse AI solver.
We showcase our methods for generating complex SAT and MILP instances utilizing generative models that mirror multifaceted structures of real-world problem.
We detail the incorporation of state-of-the-art parameter tuning algorithms which markedly elevate solver performance.
arXiv Detail & Related papers (2024-01-11T15:02:15Z) - Solving the capacitated vehicle routing problem with timing windows
using rollouts and MAX-SAT [4.873362301533824]
Vehicle routing is a well known class of NP-hard optimisation problems.
Recent work in reinforcement learning has been a promising alternative approach.
This paper proposes a hybrid approach that combines reinforcement learning, policy rollouts, and a satisfiability.
arXiv Detail & Related papers (2022-06-14T06:27:09Z) - The Machine Learning for Combinatorial Optimization Competition (ML4CO):
Results and Insights [59.93939636422896]
The ML4CO aims at improving state-of-the-art optimization solvers by replacing key components.
The competition featured three challenging tasks: finding the best feasible solution, producing the tightest optimality certificate, and giving an appropriate routing configuration.
arXiv Detail & Related papers (2022-03-04T17:06:00Z) - End-to-End Constrained Optimization Learning: A Survey [69.22203885491534]
It focuses on surveying the work on integrating solvers and optimization methods with machine learning architectures.
These approaches hold the promise to develop new hybrid machine learning and optimization methods to predict fast, approximate, structural, solutions to problems and to enable logical inference.
arXiv Detail & Related papers (2021-03-30T14:19:30Z) - A Two-stage Framework and Reinforcement Learning-based Optimization
Algorithms for Complex Scheduling Problems [54.61091936472494]
We develop a two-stage framework, in which reinforcement learning (RL) and traditional operations research (OR) algorithms are combined together.
The scheduling problem is solved in two stages, including a finite Markov decision process (MDP) and a mixed-integer programming process, respectively.
Results show that the proposed algorithms could stably and efficiently obtain satisfactory scheduling schemes for agile Earth observation satellite scheduling problems.
arXiv Detail & Related papers (2021-03-10T03:16:12Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Combining Reinforcement Learning and Constraint Programming for
Combinatorial Optimization [5.669790037378094]
The goal is to find an optimal solution among a finite set of possibilities.
Deep reinforcement learning (DRL) has shown its promise for solving NP-hard optimization problems.
constraint programming (CP) is a generic tool to solve optimization problems.
In this work, we propose a general and hybrid approach, based on DRL and CP, for solving optimization problems.
arXiv Detail & Related papers (2020-06-02T13:54:27Z) - Reinforcement Learning for Combinatorial Optimization: A Survey [12.323976053967066]
Many traditional algorithms for solving optimization problems involve using hand-crafteds that sequentially construct a solution.
Reinforcement learning (RL) proposes a good alternative to automate the search of theses by training an agent in a supervised or self-supervised manner.
arXiv Detail & Related papers (2020-03-07T16:19:45Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.