Natural Evolution Strategy for Mixed-Integer Black-Box Optimization
- URL: http://arxiv.org/abs/2304.10724v1
- Date: Fri, 21 Apr 2023 03:34:22 GMT
- Title: Natural Evolution Strategy for Mixed-Integer Black-Box Optimization
- Authors: Koki Ikeda and Isao Ono
- Abstract summary: CMA-ES w. Margin reportedly showed good performance on MI-BBO benchmark problems.
DX-NES-ICI was up to 3.7 times better than CMA-ES w. Margin in terms of a rate of finding the optimal solutions.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a natural evolution strategy (NES) for mixed-integer
black-box optimization (MI-BBO) that appears in real-world problems such as
hyperparameter optimization of machine learning and materials design. This
problem is difficult to optimize because plateaus where the values do not
change appear when the integer variables are relaxed to the continuous ones.
CMA-ES w. Margin that addresses the plateaus reportedly showed good performance
on MI-BBO benchmark problems. However, it has been observed that the search
performance of CMA-ES w. Margin deteriorates when continuous variables
contribute more to the objective function value than integer ones. In order to
address the problem of CMA-ES w. Margin, we propose Distance-weighted
eXponential Natural Evolution Strategy taking account of Implicit Constraint
and Integer (DX-NES-ICI). We compare the search performance of DX-NES-ICI with
that of CMA-ES w. Margin through numerical experiments. As a result, DX-NES-ICI
was up to 3.7 times better than CMA-ES w. Margin in terms of a rate of finding
the optimal solutions on benchmark problems where continuous variables
contribute more to the objective function value than integer ones. DX-NES-ICI
also outperformed CMA-ES w. Margin on problems where CMA-ES w. Margin
originally showed good performance.
Related papers
- CatCMA : Stochastic Optimization for Mixed-Category Problems [9.13074910982872]
Black-box optimization problems often require simultaneously optimizing different types of variables, such as continuous, integer, and categorical variables.
Several Bayesian optimization methods can deal with mixed-category black-box optimization (MC-BBO), but they suffer from a lack of scalability to high-dimensional problems and internal computational cost.
This paper proposes CatCMA, a new optimization method for MC-BBO problems.
arXiv Detail & Related papers (2024-05-16T10:11:18Z) - (1+1)-CMA-ES with Margin for Discrete and Mixed-Integer Problems [6.181926091202142]
This paper introduces the margin correction into (1+1)-CMA-ES, an elitist version of CMA-ES.
The (1+1)-CMA-ES is often advantageous for unimodal functions and can be computationally less expensive.
The numerical simulation using benchmark functions on mixed-integer, integer, and binary domains shows that (1+1)-CMA-ES with margin outperforms the CMA-ES with margin.
arXiv Detail & Related papers (2023-05-01T14:34:55Z) - Marginal Probability-Based Integer Handling for CMA-ES Tackling
Single-and Multi-Objective Mixed-Integer Black-Box Optimization [10.42820615166362]
We study the mixed-integer black-box optimization (MI-BBO) problem where continuous and integer variables should be optimized simultaneously.
We propose a simple integer handling for the CMA-ES based on lower-bounding the marginal probabilities associated with the generation of integer variables in the MGD.
arXiv Detail & Related papers (2022-12-19T06:08:20Z) - Bayesian Optimization for Macro Placement [48.55456716632735]
We develop a novel approach to macro placement using Bayesian optimization (BO) over sequence pairs.
BO is a machine learning technique that uses a probabilistic surrogate model and an acquisition function.
We demonstrate our algorithm on the fixed-outline macro placement problem with the half-perimeter wire length objective.
arXiv Detail & Related papers (2022-07-18T06:17:06Z) - CMA-ES with Margin: Lower-Bounding Marginal Probability for
Mixed-Integer Black-Box Optimization [5.237999056930947]
This study targets the mixed-integer black-box optimization (MI-BBO) problem where continuous and integer variables should be optimized simultaneously.
We propose a simple modification of the CMA-ES based on lower-bounding the marginal probabilities associated with the generation of integer variables in the MGD.
arXiv Detail & Related papers (2022-05-26T16:47:56Z) - RoMA: Robust Model Adaptation for Offline Model-based Optimization [115.02677045518692]
We consider the problem of searching an input maximizing a black-box objective function given a static dataset of input-output queries.
A popular approach to solving this problem is maintaining a proxy model that approximates the true objective function.
Here, the main challenge is how to avoid adversarially optimized inputs during the search.
arXiv Detail & Related papers (2021-10-27T05:37:12Z) - Natural Evolution Strategy for Unconstrained and Implicitly Constrained
Problems with Ridge Structure [2.512827436728378]
We propose a new natural evolution strategy for unconstrained black-box function optimization (BBFO) problems and implicitly constrained BBFO problems.
FM-NES shows better performance than DX-NES-IC on problems with ridge structure and almost the same performance as DX-NES-IC on the others.
arXiv Detail & Related papers (2021-08-21T08:01:56Z) - Robust Value Iteration for Continuous Control Tasks [99.00362538261972]
When transferring a control policy from simulation to a physical system, the policy needs to be robust to variations in the dynamics to perform well.
We present Robust Fitted Value Iteration, which uses dynamic programming to compute the optimal value function on the compact state domain.
We show that robust value is more robust compared to deep reinforcement learning algorithm and the non-robust version of the algorithm.
arXiv Detail & Related papers (2021-05-25T19:48:35Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Least Squares Regression with Markovian Data: Fundamental Limits and
Algorithms [69.45237691598774]
We study the problem of least squares linear regression where the data-points are dependent and are sampled from a Markov chain.
We establish sharp information theoretic minimax lower bounds for this problem in terms of $tau_mathsfmix$.
We propose an algorithm based on experience replay--a popular reinforcement learning technique--that achieves a significantly better error rate.
arXiv Detail & Related papers (2020-06-16T04:26:50Z) - Distributionally Robust Bayesian Optimization [121.71766171427433]
We present a novel distributionally robust Bayesian optimization algorithm (DRBO) for zeroth-order, noisy optimization.
Our algorithm provably obtains sub-linear robust regret in various settings.
We demonstrate the robust performance of our method on both synthetic and real-world benchmarks.
arXiv Detail & Related papers (2020-02-20T22:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.