Benchmarking Algorithms for Submodular Optimization Problems Using
IOHProfiler
- URL: http://arxiv.org/abs/2302.01464v1
- Date: Thu, 2 Feb 2023 23:36:23 GMT
- Title: Benchmarking Algorithms for Submodular Optimization Problems Using
IOHProfiler
- Authors: Frank Neumann, Aneta Neumann, Chao Qian, Viet Anh Do, Jacob de Nobel,
Diederick Vermetten, Saba Sadeghi Ahouei, Furong Ye, Hao Wang, Thomas B\"ack
- Abstract summary: This paper introduces a setup for benchmarking algorithms for submodular optimization problems.
The focus is on the development of iterative search algorithms with the implementation provided and integrated into IOHprofiler.
We present a range of submodular optimization problems that have been integrated into IOHprofiler and show how the setup can be used for analyzing and comparing iterative search algorithms in various settings.
- Score: 22.08617448389877
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Submodular functions play a key role in the area of optimization as they
allow to model many real-world problems that face diminishing returns.
Evolutionary algorithms have been shown to obtain strong theoretical
performance guarantees for a wide class of submodular problems under various
types of constraints while clearly outperforming standard greedy approximation
algorithms. This paper introduces a setup for benchmarking algorithms for
submodular optimization problems with the aim to provide researchers with a
framework to enhance and compare the performance of new algorithms for
submodular problems. The focus is on the development of iterative search
algorithms such as evolutionary algorithms with the implementation provided and
integrated into IOHprofiler which allows for tracking and comparing the
progress and performance of iterative search algorithms. We present a range of
submodular optimization problems that have been integrated into IOHprofiler and
show how the setup can be used for analyzing and comparing iterative search
algorithms in various settings.
Related papers
- BMR and BWR: Two simple metaphor-free optimization algorithms for solving real-life non-convex constrained and unconstrained problems [0.5755004576310334]
Two simple yet powerful optimization algorithms, named the Best-MeanRandom (BMR) and Best-Worst-Random (BWR) algorithms, are developed and presented in this paper.
arXiv Detail & Related papers (2024-07-15T18:11:47Z) - Performance Evaluation of Evolutionary Algorithms for Analog Integrated
Circuit Design Optimisation [0.0]
An automated sizing approach for analog circuits is presented in this paper.
A targeted search of the search space has been implemented using a particle generation function and a repair-bounds function.
The algorithms are tuned and modified to converge to a better optimal solution.
arXiv Detail & Related papers (2023-10-19T03:26:36Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - Hybrid ACO-CI Algorithm for Beam Design problems [0.4397520291340694]
A novel hybrid version of the Ant colony optimization (ACO) method is developed using the sample space reduction technique of the Cohort Intelligence (CI) algorithm.
The proposed work could be investigate for real world applications encompassing domains of engineering, and health care problems.
arXiv Detail & Related papers (2023-03-29T04:37:14Z) - A Simple Evolutionary Algorithm for Multi-modal Multi-objective
Optimization [0.0]
We introduce a steady-state evolutionary algorithm for solving multi-modal, multi-objective optimization problems (MMOPs)
We report its performance on 21 MMOPs from various test suites that are widely used for benchmarking using a low computational budget of 1000 function evaluations.
arXiv Detail & Related papers (2022-01-18T03:31:11Z) - A survey on multi-objective hyperparameter optimization algorithms for
Machine Learning [62.997667081978825]
This article presents a systematic survey of the literature published between 2014 and 2020 on multi-objective HPO algorithms.
We distinguish between metaheuristic-based algorithms, metamodel-based algorithms, and approaches using a mixture of both.
We also discuss the quality metrics used to compare multi-objective HPO procedures and present future research directions.
arXiv Detail & Related papers (2021-11-23T10:22:30Z) - Provably Faster Algorithms for Bilevel Optimization [54.83583213812667]
Bilevel optimization has been widely applied in many important machine learning applications.
We propose two new algorithms for bilevel optimization.
We show that both algorithms achieve the complexity of $mathcalO(epsilon-1.5)$, which outperforms all existing algorithms by the order of magnitude.
arXiv Detail & Related papers (2021-06-08T21:05:30Z) - PAMELI: A Meta-Algorithm for Computationally Expensive Multi-Objective
Optimization Problems [0.0]
The proposed algorithm is based on solving a set of surrogate problems defined by models of the real one.
Our algorithm also performs a meta-search for optimal surrogate models and navigation strategies for the optimization landscape.
arXiv Detail & Related papers (2021-03-19T11:18:03Z) - Iterative Algorithm Induced Deep-Unfolding Neural Networks: Precoding
Design for Multiuser MIMO Systems [59.804810122136345]
We propose a framework for deep-unfolding, where a general form of iterative algorithm induced deep-unfolding neural network (IAIDNN) is developed.
An efficient IAIDNN based on the structure of the classic weighted minimum mean-square error (WMMSE) iterative algorithm is developed.
We show that the proposed IAIDNN efficiently achieves the performance of the iterative WMMSE algorithm with reduced computational complexity.
arXiv Detail & Related papers (2020-06-15T02:57:57Z) - Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization [71.03797261151605]
Adaptivity is an important yet under-studied property in modern optimization theory.
Our algorithm is proved to achieve the best-available convergence for non-PL objectives simultaneously while outperforming existing algorithms for PL objectives.
arXiv Detail & Related papers (2020-02-13T05:42:27Z) - Extreme Algorithm Selection With Dyadic Feature Representation [78.13985819417974]
We propose the setting of extreme algorithm selection (XAS) where we consider fixed sets of thousands of candidate algorithms.
We assess the applicability of state-of-the-art AS techniques to the XAS setting and propose approaches leveraging a dyadic feature representation.
arXiv Detail & Related papers (2020-01-29T09:40:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.