Using Automated Algorithm Configuration for Parameter Control
- URL: http://arxiv.org/abs/2302.12334v2
- Date: Mon, 14 Aug 2023 12:37:30 GMT
- Title: Using Automated Algorithm Configuration for Parameter Control
- Authors: Deyao Chen, Maxim Buzdalov, Carola Doerr, Nguyen Dang
- Abstract summary: Dynamic Algorithm configuration (DAC) tackles the question of how to automatically learn policies to control parameters of algorithms in a data-driven fashion.
We propose a new DAC benchmark the controlling of the key parameter $lambda$ in the $(lambda,lambda)$Genetic Algorithm for solving OneMax problems.
Our approach is able to consistently outperform the default parameter control policy of the benchmark derived from previous theoretical work on sufficiently large problem sizes.
- Score: 0.7742297876120562
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic Algorithm Configuration (DAC) tackles the question of how to
automatically learn policies to control parameters of algorithms in a
data-driven fashion. This question has received considerable attention from the
evolutionary community in recent years. Having a good benchmark collection to
gain structural understanding on the effectiveness and limitations of different
solution methods for DAC is therefore strongly desirable. Following recent work
on proposing DAC benchmarks with well-understood theoretical properties and
ground truth information, in this work, we suggest as a new DAC benchmark the
controlling of the key parameter $\lambda$ in the
$(1+(\lambda,\lambda))$~Genetic Algorithm for solving OneMax problems. We
conduct a study on how to solve the DAC problem via the use of (static)
automated algorithm configuration on the benchmark, and propose techniques to
significantly improve the performance of the approach. Our approach is able to
consistently outperform the default parameter control policy of the benchmark
derived from previous theoretical work on sufficiently large problem sizes. We
also present new findings on the landscape of the parameter-control search
policies and propose methods to compute stronger baselines for the benchmark
via numerical approximations of the true optimal policies.
Related papers
- Deep Reinforcement Learning for Online Optimal Execution Strategies [49.1574468325115]
This paper tackles the challenge of learning non-Markovian optimal execution strategies in dynamic financial markets.
We introduce a novel actor-critic algorithm based on Deep Deterministic Policy Gradient (DDPG)
We show that our algorithm successfully approximates the optimal execution strategy.
arXiv Detail & Related papers (2024-10-17T12:38:08Z) - Automated Dynamic Algorithm Configuration [39.39845379026921]
The performance of an algorithm often critically depends on its parameter configuration.
It has been shown that some algorithm parameters are best adjusted dynamically during execution.
A promising alternative is to automatically learn such dynamic parameter adaptation policies from data.
arXiv Detail & Related papers (2022-05-27T10:30:25Z) - Theory-inspired Parameter Control Benchmarks for Dynamic Algorithm
Configuration [32.055812915031666]
We show how to compute optimal parameter portfolios of a given size.
We extend this benchmark by analyzing optimal control policies that can select the parameters only from a given portfolio of possible values.
We demonstrate the usefulness of our benchmarks by analyzing the behavior of the DDQN reinforcement learning approach for dynamic algorithm configuration.
arXiv Detail & Related papers (2022-02-07T15:00:30Z) - Instance-Dependent Confidence and Early Stopping for Reinforcement
Learning [99.57168572237421]
Various algorithms for reinforcement learning (RL) exhibit dramatic variation in their convergence rates as a function of problem structure.
This research provides guarantees that explain textitex post the performance differences observed.
A natural next step is to convert these theoretical guarantees into guidelines that are useful in practice.
arXiv Detail & Related papers (2022-01-21T04:25:35Z) - Offline RL Without Off-Policy Evaluation [49.11859771578969]
We show that simply doing one step of constrained/regularized policy improvement using an on-policy Q estimate of the behavior policy performs surprisingly well.
This one-step algorithm beats the previously reported results of iterative algorithms on a large portion of the D4RL benchmark.
arXiv Detail & Related papers (2021-06-16T16:04:26Z) - DACBench: A Benchmark Library for Dynamic Algorithm Configuration [30.217571636151295]
We propose DACBench, a benchmark library that seeks to collect and standardize existing DAC benchmarks from different AI domains.
To show the potential, broad applicability and challenges of DAC, we explore how a set of six initial benchmarks compare in several dimensions of difficulty.
arXiv Detail & Related papers (2021-05-18T14:16:51Z) - Logistic Q-Learning [87.00813469969167]
We propose a new reinforcement learning algorithm derived from a regularized linear-programming formulation of optimal control in MDPs.
The main feature of our algorithm is a convex loss function for policy evaluation that serves as a theoretically sound alternative to the widely used squared Bellman error.
arXiv Detail & Related papers (2020-10-21T17:14:31Z) - Variance-Reduced Off-Policy Memory-Efficient Policy Search [61.23789485979057]
Off-policy policy optimization is a challenging problem in reinforcement learning.
Off-policy algorithms are memory-efficient and capable of learning from off-policy samples.
arXiv Detail & Related papers (2020-09-14T16:22:46Z) - Robust Reinforcement Learning using Least Squares Policy Iteration with
Provable Performance Guarantees [3.8073142980733]
This paper addresses the problem of model-free reinforcement learning for Robust Markov Decision Process (RMDP) with large state spaces.
We first propose the Robust Least Squares Policy Evaluation algorithm, which is a multi-step online model-free learning algorithm for policy evaluation.
We then propose Robust Least Squares Policy Iteration (RLSPI) algorithm for learning the optimal robust policy.
arXiv Detail & Related papers (2020-06-20T16:26:50Z) - MATE: A Model-based Algorithm Tuning Engine [2.4693304175649304]
We introduce a Model-based Algorithm Turning Engine, namely MATE, where the parameters of an algorithm are represented as expressions of the features of a target optimisation problem.
We formulate the problem of finding the relationships between the parameters and the problem features as a symbolic regression problem and we use genetic programming to extract these expressions.
For the evaluation, we apply our approach to configuration of the (1+1) EA and RLS algorithms for the OneMax, LeadingOnes, BinValue and Jump optimisation problems.
arXiv Detail & Related papers (2020-04-27T12:50:48Z) - Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization [71.03797261151605]
Adaptivity is an important yet under-studied property in modern optimization theory.
Our algorithm is proved to achieve the best-available convergence for non-PL objectives simultaneously while outperforming existing algorithms for PL objectives.
arXiv Detail & Related papers (2020-02-13T05:42:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.