Efficient Incremental Modelling and Solving
- URL: http://arxiv.org/abs/2009.11111v1
- Date: Wed, 23 Sep 2020 12:40:23 GMT
- Title: Efficient Incremental Modelling and Solving
- Authors: G\"okberk Ko\c{c}ak, \"Ozg\"ur Akg\"un, Nguyen Dang, Ian Miguel
- Abstract summary: A standard approach to solving AI planning problems is to incrementally extend the planning horizon and solve the problem of trying to find a plan of a particular length.
The contribution of this work is to enable a native interaction between SAT solvers and the automated modelling system Savile Row.
- Score: 1.6172800007896284
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In various scenarios, a single phase of modelling and solving is either not
sufficient or not feasible to solve the problem at hand. A standard approach to
solving AI planning problems, for example, is to incrementally extend the
planning horizon and solve the problem of trying to find a plan of a particular
length. Indeed, any optimization problem can be solved as a sequence of
decision problems in which the objective value is incrementally updated.
Another example is constraint dominance programming (CDP), in which search is
organized into a sequence of levels. The contribution of this work is to enable
a native interaction between SAT solvers and the automated modelling system
Savile Row to support efficient incremental modelling and solving. This allows
adding new decision variables, posting new constraints and removing existing
constraints (via assumptions) between incremental steps. Two additional
benefits of the native coupling of modelling and solving are the ability to
retain learned information between SAT solver calls and to enable SAT
assumptions, further improving flexibility and efficiency. Experiments on one
optimisation problem and five pattern mining tasks demonstrate that the native
interaction between the modelling system and SAT solver consistently improves
performance significantly.
Related papers
- SEGO: Sequential Subgoal Optimization for Mathematical Problem-Solving [64.38649623473626]
Large Language Models (LLMs) have driven substantial progress in artificial intelligence.
We propose a novel framework called textbfSEquential subtextbfGoal textbfOptimization (SEGO) to enhance LLMs' ability to solve mathematical problems.
arXiv Detail & Related papers (2023-10-19T17:56:40Z) - On data-driven chance constraint learning for mixed-integer optimization
problems [0.0]
We develop a Chance Constraint Learning (CCL) methodology with a focus on mixed-integer linear optimization problems.
CCL makes use of linearizable machine learning models to estimate conditional quantiles of the learned variables.
An open-access software has been developed to be used by practitioners.
arXiv Detail & Related papers (2022-07-08T11:54:39Z) - Efficient lifting of symmetry breaking constraints for complex
combinatorial problems [9.156939957189502]
This work extends the learning framework and implementation of a model-based approach for Answer Set Programming.
In particular, we incorporate a new conflict analysis algorithm in the Inductive Logic Programming system ILASP.
arXiv Detail & Related papers (2022-05-14T20:42:13Z) - Neural Stochastic Dual Dynamic Programming [99.80617899593526]
We introduce a trainable neural model that learns to map problem instances to a piece-wise linear value function.
$nu$-SDDP can significantly reduce problem solving cost without sacrificing solution quality.
arXiv Detail & Related papers (2021-12-01T22:55:23Z) - Transformer-based Machine Learning for Fast SAT Solvers and Logic
Synthesis [63.53283025435107]
CNF-based SAT and MaxSAT solvers are central to logic synthesis and verification systems.
In this work, we propose a one-shot model derived from the Transformer architecture to solve the MaxSAT problem.
arXiv Detail & Related papers (2021-07-15T04:47:35Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Simplified Swarm Optimization for Bi-Objection Active Reliability
Redundancy Allocation Problems [1.5990720051907859]
The reliability redundancy allocation problem (RRAP) is a well-known problem in system design, development, and management.
In this study, a bi-objective RRAP is formulated by changing the cost constraint as a new goal.
To solve the proposed problem, a new simplified swarm optimization (SSO) with a penalty function, a real one-type solution structure, a number-based self-adaptive new update mechanism, a constrained non-dominated solution selection, and a new pBest replacement policy is developed.
arXiv Detail & Related papers (2020-06-17T13:15:44Z) - Consistent Second-Order Conic Integer Programming for Learning Bayesian
Networks [2.7473982588529653]
We study the problem of learning the sparse DAG structure of a BN from continuous observational data.
The optimal solution to this mathematical program is known to have desirable statistical properties under certain conditions.
We propose a concrete early stopping criterion to terminate the branch-and-bound process in order to obtain a near-optimal solution.
arXiv Detail & Related papers (2020-05-29T00:13:15Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.