Strong Structural Bounds for MaxSAT: The Fine Details of Using Neuromorphic and Quantum Hardware Accelerators
- URL: http://arxiv.org/abs/2412.10289v1
- Date: Fri, 13 Dec 2024 17:18:08 GMT
- Title: Strong Structural Bounds for MaxSAT: The Fine Details of Using Neuromorphic and Quantum Hardware Accelerators
- Authors: Max Bannach, Jai Grover, Markus Hecher,
- Abstract summary: Hardware accelerators like quantum annealers or neuromorphic chips are capable of finding the ground state of a Hamiltonian.
Previous studies were only concerned with the size of the encodings rather than with syntactic or structural properties.
We establish structure-aware reductions between MaxSAT, Max2SAT, and the quadratic unconstrained binary optimization problem (QUBO) that underlies such hardware accelerators.
- Score: 16.324368802296526
- License:
- Abstract: Hardware accelerators like quantum annealers or neuromorphic chips are capable of finding the ground state of a Hamiltonian. A promising route in utilizing these devices is via methods from automated reasoning: The problem at hand is first encoded into MaxSAT; then MaxSAT is reduced to Max2SAT; and finally, Max2SAT is translated into a Hamiltonian. It was observed that different encodings can dramatically affect the efficiency of the hardware accelerators. Yet, previous studies were only concerned with the size of the encodings rather than with syntactic or structural properties. We establish structure-aware reductions between MaxSAT, Max2SAT, and the quadratic unconstrained binary optimization problem (QUBO) that underlies such hardware accelerators. All these problems turn out to be equivalent under linear-time, treewidth-preserving reductions. As a consequence, we obtain tight lower bounds under ETH and SETH for Max2SAT and QUBO, as well as a new time-optimal fixed-parameter algorithm for QUBO. While our results are tight up to a constant additive factor for the primal treewidth, we require a constant multiplicative factor for the incidence treewidth. To close the emerging gap, we supplement our results with novel time-optimal algorithms for fragments of MaxSAT based on model counting.
Related papers
- MaxSAT decoders for arbitrary CSS codes [2.3895981099137535]
We show how to map quantum maximum likelihood problem of CSS codes of arbitrary geometry and parity check weight into MaxSAT problems.
Our decoder can be further parallelised or implemented on ASICs and FPGAs, promising potential further speedups of several orders of magnitude.
arXiv Detail & Related papers (2024-10-02T15:45:05Z) - Solving MaxSAT with Matrix Multiplication [15.349236934751897]
We propose an algorithm for Maximum Satisfiability (MaxSAT) specifically designed to run on neural network accelerators such as GPUs and TPUs.
Our approach, which we term RbmSAT, is a new design point in the algorithm-hardware co-design space for MaxSAT.
We present timed results on a subset of problem instances from the annual MaxSAT Evaluations Incomplete Un Track for the years 2018 to 2021.
arXiv Detail & Related papers (2023-11-01T14:46:46Z) - Improving Dual-Encoder Training through Dynamic Indexes for Negative
Mining [61.09807522366773]
We introduce an algorithm that approximates the softmax with provable bounds and that dynamically maintains the tree.
In our study on datasets with over twenty million targets, our approach cuts error by half in relation to oracle brute-force negative mining.
arXiv Detail & Related papers (2023-03-27T15:18:32Z) - Towards Tackling MaxSAT by Combining Nested Monte Carlo with Local
Search [10.70006528984961]
We introduce two algorithmic variations over UCTMAXSAT.
First, a nesting of the tree search inspired by the Nested Monte Carlo Search algorithm is effective on most instance types in the benchmark.
Second, we observe that using a static flip limit in SLS, the ideal budget depends heavily on the instance size and we propose to set it dynamically.
arXiv Detail & Related papers (2023-02-26T03:37:26Z) - Softmax-free Linear Transformers [90.83157268265654]
Vision transformers (ViTs) have pushed the state-of-the-art for visual perception tasks.
Existing methods are either theoretically flawed or empirically ineffective for visual recognition.
We propose a family of Softmax-Free Transformers (SOFT)
arXiv Detail & Related papers (2022-07-05T03:08:27Z) - DPMS: An ADD-Based Symbolic Approach for Generalized MaxSAT Solving [45.21499915442282]
We propose a novel dynamic-programming approach for solving generalized MaxSAT problems with hybrid constraints.
Our versatile framework admits many generalizations of CNF-MaxSAT, such as MaxSAT, Min-MaxSAT, and MinSAT with hybrid constraints.
Empirical results indicate that DPMS is able to solve certain problems quickly, where other algorithms based on various techniques all fail.
arXiv Detail & Related papers (2022-05-08T00:31:53Z) - Transformer-based Machine Learning for Fast SAT Solvers and Logic
Synthesis [63.53283025435107]
CNF-based SAT and MaxSAT solvers are central to logic synthesis and verification systems.
In this work, we propose a one-shot model derived from the Transformer architecture to solve the MaxSAT problem.
arXiv Detail & Related papers (2021-07-15T04:47:35Z) - Incomplete MaxSAT Approaches for Combinatorial Testing [0.0]
We present a Satisfiability (SAT)-based approach for building Mixed Covering Arrays with Constraints of minimum length.
This problem is central in Combinatorial Testing for the detection of system failures.
arXiv Detail & Related papers (2021-05-26T14:00:56Z) - Improved Branch and Bound for Neural Network Verification via Lagrangian
Decomposition [161.09660864941603]
We improve the scalability of Branch and Bound (BaB) algorithms for formally proving input-output properties of neural networks.
We present a novel activation-based branching strategy and a BaB framework, named Branch and Dual Network Bound (BaDNB)
BaDNB outperforms previous complete verification systems by a large margin, cutting average verification times by factors up to 50 on adversarial properties.
arXiv Detail & Related papers (2021-04-14T09:22:42Z) - Lower Bounds and Accelerated Algorithms for Bilevel Optimization [62.192297758346484]
Bilevel optimization has recently attracted growing interests due to its wide applications in modern machine learning problems.
We show that our results are more challenging than that of minimax applications.
arXiv Detail & Related papers (2021-02-07T21:46:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.