BOiLS: Bayesian Optimisation for Logic Synthesis
- URL: http://arxiv.org/abs/2111.06178v1
- Date: Thu, 11 Nov 2021 12:44:38 GMT
- Title: BOiLS: Bayesian Optimisation for Logic Synthesis
- Authors: Antoine Grosnit, Cedric Malherbe, Rasul Tutunov, Xingchen Wan, Jun
Wang, Haitham Bou Ammar
- Abstract summary: We propose BOiLS, the first algorithm adapting modern Bayesian optimisation to navigate the space of synthesis operations.
We demonstrate BOiLS's superior performance compared to state-of-the-art in terms of both sample efficiency and QoR values.
- Score: 10.981155046738126
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optimising the quality-of-results (QoR) of circuits during logic synthesis is
a formidable challenge necessitating the exploration of exponentially sized
search spaces. While expert-designed operations aid in uncovering effective
sequences, the increase in complexity of logic circuits favours automated
procedures. Inspired by the successes of machine learning, researchers adapted
deep learning and reinforcement learning to logic synthesis applications.
However successful, those techniques suffer from high sample complexities
preventing widespread adoption. To enable efficient and scalable solutions, we
propose BOiLS, the first algorithm adapting modern Bayesian optimisation to
navigate the space of synthesis operations. BOiLS requires no human
intervention and effectively trades-off exploration versus exploitation through
novel Gaussian process kernels and trust-region constrained acquisitions. In a
set of experiments on EPFL benchmarks, we demonstrate BOiLS's superior
performance compared to state-of-the-art in terms of both sample efficiency and
QoR values.
Related papers
- EVOLvE: Evaluating and Optimizing LLMs For Exploration [76.66831821738927]
Large language models (LLMs) remain under-studied in scenarios requiring optimal decision-making under uncertainty.
We measure LLMs' (in)ability to make optimal decisions in bandits, a state-less reinforcement learning setting relevant to many applications.
Motivated by the existence of optimal exploration algorithms, we propose efficient ways to integrate this algorithmic knowledge into LLMs.
arXiv Detail & Related papers (2024-10-08T17:54:03Z) - OPTDTALS: Approximate Logic Synthesis via Optimal Decision Trees Approach [9.081146426124482]
Approximate Logic Synthesis (ALS) aims to reduce circuit complexity by sacrificing correctness.
We propose a new ALS methodology realizing approximation via learning optimal decision trees in empirical accuracy.
arXiv Detail & Related papers (2024-08-22T11:23:58Z) - Retrieval-Guided Reinforcement Learning for Boolean Circuit Minimization [23.075466444266528]
This study conducts a thorough examination of learning and search techniques for logic synthesis.
We present ABC-RL, a meticulously tuned $alpha$ parameter that adeptly adjusts recommendations from pre-trained agents during the search process.
Our findings showcase substantial enhancements in the Quality-of-result (QoR) of synthesized circuits, boasting improvements of up to 24.8% compared to state-of-the-art techniques.
arXiv Detail & Related papers (2024-01-22T18:46:30Z) - Can LLMs Configure Software Tools [0.76146285961466]
In software engineering, the meticulous configuration of software tools is crucial in ensuring optimal performance within intricate systems.
In this study, we embark on an exploration of leveraging Large-Language Models (LLMs) to streamline the software configuration process.
Our work presents a novel approach that employs LLMs, such as Chat-GPT, to identify starting conditions and narrow down the search space, improving configuration efficiency.
arXiv Detail & Related papers (2023-12-11T05:03:02Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - INVICTUS: Optimizing Boolean Logic Circuit Synthesis via Synergistic
Learning and Search [18.558280701880136]
State-of-the-art logic synthesis algorithms have a large number of logic minimizations.
INVICTUS generates a sequence of logic minimizations based on a training dataset of previously seen designs.
arXiv Detail & Related papers (2023-05-22T15:50:42Z) - Representation Learning with Multi-Step Inverse Kinematics: An Efficient
and Optimal Approach to Rich-Observation RL [106.82295532402335]
Existing reinforcement learning algorithms suffer from computational intractability, strong statistical assumptions, and suboptimal sample complexity.
We provide the first computationally efficient algorithm that attains rate-optimal sample complexity with respect to the desired accuracy level.
Our algorithm, MusIK, combines systematic exploration with representation learning based on multi-step inverse kinematics.
arXiv Detail & Related papers (2023-04-12T14:51:47Z) - Towards Automated Imbalanced Learning with Deep Hierarchical
Reinforcement Learning [57.163525407022966]
Imbalanced learning is a fundamental challenge in data mining, where there is a disproportionate ratio of training samples in each class.
Over-sampling is an effective technique to tackle imbalanced learning through generating synthetic samples for the minority class.
We propose AutoSMOTE, an automated over-sampling algorithm that can jointly optimize different levels of decisions.
arXiv Detail & Related papers (2022-08-26T04:28:01Z) - Learning to Schedule Heuristics for the Simultaneous Stochastic
Optimization of Mining Complexes [2.538209532048867]
The proposed learn-to-perturb (L2P) hyper-heuristic is a multi-neighborhood simulated annealing algorithm.
L2P is tested on several real-world mining complexes, with an emphasis on efficiency, robustness, and generalization capacity.
Results show a reduction in the number of iterations by 30-50% and in the computational time by 30-45%.
arXiv Detail & Related papers (2022-02-25T18:20:14Z) - Fast Distributionally Robust Learning with Variance Reduced Min-Max
Optimization [85.84019017587477]
Distributionally robust supervised learning is emerging as a key paradigm for building reliable machine learning systems for real-world applications.
Existing algorithms for solving Wasserstein DRSL involve solving complex subproblems or fail to make use of gradients.
We revisit Wasserstein DRSL through the lens of min-max optimization and derive scalable and efficiently implementable extra-gradient algorithms.
arXiv Detail & Related papers (2021-04-27T16:56:09Z) - Parallelization Techniques for Verifying Neural Networks [52.917845265248744]
We introduce an algorithm based on the verification problem in an iterative manner and explore two partitioning strategies.
We also introduce a highly parallelizable pre-processing algorithm that uses the neuron activation phases to simplify the neural network verification problems.
arXiv Detail & Related papers (2020-04-17T20:21:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.