BandMaxSAT: A Local Search MaxSAT Solver with Multi-armed Bandit
- URL: http://arxiv.org/abs/2201.05544v1
- Date: Fri, 14 Jan 2022 16:32:39 GMT
- Title: BandMaxSAT: A Local Search MaxSAT Solver with Multi-armed Bandit
- Authors: Jiongzhi Zheng and Kun He and Jianrong Zhou and Yan Jin and Chu-min Li
and Felip Manya
- Abstract summary: We propose a local search algorithm called BandMaxSAT, that applies a multi-armed bandit to guide the search direction.
Extensive experiments demonstrate that BandMaxSAT significantly outperforms the state-of-the-art (W)PMS local search algorithm SATLike3.0.
The resulting solver BandMaxSAT-c also outperforms some of the best state-of-the-art complete (W)PMS solvers, including SATLike-c, Loandra and TT-Open-WBO-Inc.
- Score: 16.371916145216737
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We address Partial MaxSAT (PMS) and Weighted PMS (WPMS), two practical
generalizations of the MaxSAT problem, and propose a local search algorithm
called BandMaxSAT, that applies a multi-armed bandit to guide the search
direction, for these problems. The bandit in our method is associated with all
the soft clauses in the input (W)PMS instance. Each arm corresponds to a soft
clause. The bandit model can help BandMaxSAT to select a good direction to
escape from local optima by selecting a soft clause to be satisfied in the
current step, that is, selecting an arm to be pulled. We further propose an
initialization method for (W)PMS that prioritizes both unit and binary clauses
when producing the initial solutions. Extensive experiments demonstrate that
BandMaxSAT significantly outperforms the state-of-the-art (W)PMS local search
algorithm SATLike3.0. Specifically, the number of instances in which BandMaxSAT
obtains better results is about twice that obtained by SATLike3.0. We further
combine BandMaxSAT with the complete solver TT-Open-WBO-Inc. The resulting
solver BandMaxSAT-c also outperforms some of the best state-of-the-art complete
(W)PMS solvers, including SATLike-c, Loandra and TT-Open-WBO-Inc.
Related papers
- Rethinking the Soft Conflict Pseudo Boolean Constraint on MaxSAT Local
Search Solvers [20.866863965121013]
MaxSAT is an optimization version of the famous NP-complete Satisfiability problem (SAT)
We propose a new local search algorithm called SPB-MaxSAT that provides new perspectives for clause weighting on MaxSAT local search solvers.
arXiv Detail & Related papers (2024-01-19T09:59:02Z) - Combinatorial Stochastic-Greedy Bandit [79.1700188160944]
We propose a novelgreedy bandit (SGB) algorithm for multi-armed bandit problems when no extra information other than the joint reward of the selected set of $n$ arms at each time $tin [T]$ is observed.
SGB adopts an optimized-explore-then-commit approach and is specifically designed for scenarios with a large set of base arms.
arXiv Detail & Related papers (2023-12-13T11:08:25Z) - Fixed-Budget Real-Valued Combinatorial Pure Exploration of Multi-Armed
Bandit [65.268245109828]
We first introduce the Combinatorial Successive Asign algorithm, which is the first algorithm that can identify the best action even when the size of the action class is exponentially large.
We show that the upper bound of the probability of error of the CSA algorithm matches a lower bound up to a logarithmic factor in the exponent.
We experimentally compare the algorithms with previous methods and show that our algorithm performs better.
arXiv Detail & Related papers (2023-10-24T09:47:32Z) - Towards Tackling MaxSAT by Combining Nested Monte Carlo with Local
Search [10.70006528984961]
We introduce two algorithmic variations over UCTMAXSAT.
First, a nesting of the tree search inspired by the Nested Monte Carlo Search algorithm is effective on most instance types in the benchmark.
Second, we observe that using a static flip limit in SLS, the ideal budget depends heavily on the instance size and we propose to set it dynamically.
arXiv Detail & Related papers (2023-02-26T03:37:26Z) - Incorporating Multi-armed Bandit with Local Search for MaxSAT [16.371916145216737]
Two generalizations of the MaxSAT problem: Partial MaxSAT (PMS) and Weighted PMS (WPMS)
We propose a local search algorithm for these problems, called BandHS, which applies two multi-armed bandits to guide the search directions when escaping local optima.
These two bandits can improve the algorithm's search ability in both feasible and infeasible solution spaces.
arXiv Detail & Related papers (2022-11-29T08:19:26Z) - DPMS: An ADD-Based Symbolic Approach for Generalized MaxSAT Solving [45.21499915442282]
We propose a novel dynamic-programming approach for solving generalized MaxSAT problems with hybrid constraints.
Our versatile framework admits many generalizations of CNF-MaxSAT, such as MaxSAT, Min-MaxSAT, and MinSAT with hybrid constraints.
Empirical results indicate that DPMS is able to solve certain problems quickly, where other algorithms based on various techniques all fail.
arXiv Detail & Related papers (2022-05-08T00:31:53Z) - Transformer-based Machine Learning for Fast SAT Solvers and Logic
Synthesis [63.53283025435107]
CNF-based SAT and MaxSAT solvers are central to logic synthesis and verification systems.
In this work, we propose a one-shot model derived from the Transformer architecture to solve the MaxSAT problem.
arXiv Detail & Related papers (2021-07-15T04:47:35Z) - Incomplete MaxSAT Approaches for Combinatorial Testing [0.0]
We present a Satisfiability (SAT)-based approach for building Mixed Covering Arrays with Constraints of minimum length.
This problem is central in Combinatorial Testing for the detection of system failures.
arXiv Detail & Related papers (2021-05-26T14:00:56Z) - Upper Confidence Bounds for Combining Stochastic Bandits [52.10197476419621]
We provide a simple method to combine bandit algorithms.
Our approach is based on a "meta-UCB" procedure that treats each of $N$ individual bandit algorithms as arms in a higher-level $N$-armed bandit problem.
arXiv Detail & Related papers (2020-12-24T05:36:29Z) - MOTS: Minimax Optimal Thompson Sampling [89.2370817955411]
It has remained an open problem whether Thompson sampling can match the minimax lower bound $Omega(sqrtKT)$ for $K$-armed bandit problems.
We propose a variant of Thompson sampling called MOTS that adaptively clips the sampling instance of the chosen arm at each time step.
We prove that this simple variant of Thompson sampling achieves the minimax optimal regret bound $O(sqrtKT)$ for finite time horizon $T$, as well as the optimal regret bound for Gaussian rewards when $T$ approaches infinity.
arXiv Detail & Related papers (2020-03-03T21:24:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.