Continuous surrogate-based optimization algorithms are well-suited for
expensive discrete problems
- URL: http://arxiv.org/abs/2011.03431v2
- Date: Mon, 30 Nov 2020 19:21:54 GMT
- Title: Continuous surrogate-based optimization algorithms are well-suited for
expensive discrete problems
- Authors: Rickard Karlsson, Laurens Bliek, Sicco Verwer, Mathijs de Weerdt
- Abstract summary: We present empirical evidence showing that the use of continuous surrogate models displays competitive performance against state-of-the-art discrete surrogate-based methods.
Our experiments on different discrete structures and time constraints also give more insight into which algorithms work well on which type of problem.
- Score: 9.655888042539495
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One method to solve expensive black-box optimization problems is to use a
surrogate model that approximates the objective based on previous observed
evaluations. The surrogate, which is cheaper to evaluate, is optimized instead
to find an approximate solution to the original problem. In the case of
discrete problems, recent research has revolved around surrogate models that
are specifically constructed to deal with discrete structures. A main
motivation is that literature considers continuous methods, such as Bayesian
optimization with Gaussian processes as the surrogate, to be sub-optimal
(especially in higher dimensions) because they ignore the discrete structure
by, e.g., rounding off real-valued solutions to integers. However, we claim
that this is not true. In fact, we present empirical evidence showing that the
use of continuous surrogate models displays competitive performance on a set of
high-dimensional discrete benchmark problems, including a real-life
application, against state-of-the-art discrete surrogate-based methods. Our
experiments on different discrete structures and time constraints also give
more insight into which algorithms work well on which type of problem.
Related papers
- An Efficient Approach for Solving Expensive Constrained Multiobjective Optimization Problems [0.0]
An efficient probabilistic selection based constrained multi-objective EA is proposed, referred to as PSCMOEA.
It comprises novel elements such as (a) an adaptive search bound identification scheme based on the feasibility and convergence status of evaluated solutions.
Numerical experiments are conducted on an extensive range of challenging constrained problems using low evaluation budgets to simulate ECMOPs.
arXiv Detail & Related papers (2024-05-22T02:32:58Z) - Discretize Relaxed Solution of Spectral Clustering via a Non-Heuristic
Algorithm [77.53604156112144]
We develop a first-order term to bridge the original problem and discretization algorithm.
Since the non-heuristic method is aware of the original graph cut problem, the final discrete solution is more reliable.
arXiv Detail & Related papers (2023-10-19T13:57:38Z) - Backpropagation of Unrolled Solvers with Folded Optimization [55.04219793298687]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
One typical strategy is algorithm unrolling, which relies on automatic differentiation through the operations of an iterative solver.
This paper provides theoretical insights into the backward pass of unrolled optimization, leading to a system for generating efficiently solvable analytical models of backpropagation.
arXiv Detail & Related papers (2023-01-28T01:50:42Z) - Exploring the Algorithm-Dependent Generalization of AUPRC Optimization
with List Stability [107.65337427333064]
optimization of the Area Under the Precision-Recall Curve (AUPRC) is a crucial problem for machine learning.
In this work, we present the first trial in the single-dependent generalization of AUPRC optimization.
Experiments on three image retrieval datasets on speak to the effectiveness and soundness of our framework.
arXiv Detail & Related papers (2022-09-27T09:06:37Z) - Uncertainty-Aware Search Framework for Multi-Objective Bayesian
Optimization [40.40632890861706]
We consider the problem of multi-objective (MO) blackbox optimization using expensive function evaluations.
We propose a novel uncertainty-aware search framework referred to as USeMO to efficiently select the sequence of inputs for evaluation.
arXiv Detail & Related papers (2022-04-12T16:50:48Z) - Learning Proximal Operators to Discover Multiple Optima [66.98045013486794]
We present an end-to-end method to learn the proximal operator across non-family problems.
We show that for weakly-ized objectives and under mild conditions, the method converges globally.
arXiv Detail & Related papers (2022-01-28T05:53:28Z) - Outlier-Robust Sparse Estimation via Non-Convex Optimization [73.18654719887205]
We explore the connection between high-dimensional statistics and non-robust optimization in the presence of sparsity constraints.
We develop novel and simple optimization formulations for these problems.
As a corollary, we obtain that any first-order method that efficiently converges to station yields an efficient algorithm for these tasks.
arXiv Detail & Related papers (2021-09-23T17:38:24Z) - Batched Data-Driven Evolutionary Multi-Objective Optimization Based on
Manifold Interpolation [6.560512252982714]
We propose a framework for implementing batched data-driven evolutionary multi-objective optimization.
It is so general that any off-the-shelf evolutionary multi-objective optimization algorithms can be applied in a plug-in manner.
Our proposed framework is featured with a faster convergence and a stronger resilience to various PF shapes.
arXiv Detail & Related papers (2021-09-12T23:54:26Z) - Joint Continuous and Discrete Model Selection via Submodularity [1.332560004325655]
In model selection problems for machine learning, the desire for a well-performing model with meaningful structure is typically expressed through a regularized optimization problem.
In many scenarios, however, numerically meaningful structure is specified in some discrete space leading to difficult non optimization problems.
We show how simple continuous or discrete constraints can also be handled for certain problem classes, motivated by robust optimization.
arXiv Detail & Related papers (2021-02-17T21:14:47Z) - Population-Based Methods: PARTICLE SWARM OPTIMIZATION -- Development of
a General-Purpose Optimizer and Applications [0.0]
This thesis is concerned with continuous, static, and single-objective optimization problems subject to inequality constraints.
The particle swarm optimization paradigm was inspired by previous simulations of the cooperative behaviour observed in social beings.
arXiv Detail & Related papers (2021-01-25T09:36:25Z) - Robust, Accurate Stochastic Optimization for Variational Inference [68.83746081733464]
We show that common optimization methods lead to poor variational approximations if the problem is moderately large.
Motivated by these findings, we develop a more robust and accurate optimization framework by viewing the underlying algorithm as producing a Markov chain.
arXiv Detail & Related papers (2020-09-01T19:12:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.