An Extended Jump Function Benchmark for the Analysis of Randomized
Search Heuristics
- URL: http://arxiv.org/abs/2105.03090v1
- Date: Fri, 7 May 2021 07:21:10 GMT
- Title: An Extended Jump Function Benchmark for the Analysis of Randomized
Search Heuristics
- Authors: Henry Bambury, Antoine Bultel, Benjamin Doerr
- Abstract summary: We propose an extended class $textscJump_k,delta$ of jump functions that contain a valley of low fitness of width.
We show that the new class allows experiments with wider fitness valleys, especially when they lie further away from the global optimum.
- Score: 4.38301148531795
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Jump functions are the most studied non-unimodal benchmark in the theory of
randomized search heuristics, in particular, evolutionary algorithms (EAs).
They have significantly improved our understanding of how EAs escape from local
optima. However, their particular structure -- to leave the local optimum one
can only jump directly to the global optimum -- raises the question of how
representative such results are.
For this reason, we propose an extended class $\textsc{Jump}_{k,\delta}$ of
jump functions that contain a valley of low fitness of width $\delta$ starting
at distance $k$ from the global optimum. We prove that several previous results
extend to this more general class: for all $k = o(n^{1/3})$ and $\delta < k$,
the optimal mutation rate for the $(1+1)$~EA is $\frac{\delta}{n}$, and the
fast $(1+1)$~EA runs faster than the classical $(1+1)$~EA by a factor
super-exponential in $\delta$. However, we also observe that some known results
do not generalize: the randomized local search algorithm with stagnation
detection, which is faster than the fast $(1+1)$~EA by a factor polynomial in
$k$ on $\textsc{Jump}_k$, is slower by a factor polynomial in $n$ on some
$\textsc{Jump}_{k,\delta}$ instances.
Computationally, the new class allows experiments with wider fitness valleys,
especially when they lie further away from the global optimum.
Related papers
- Hyper-Heuristics Can Profit From Global Variation Operators [12.774575491521926]
We show that the Move Acceptance Hyper-Heuristic (MAHH) leaves the local optimum of the multimodal CLIFF benchmark with remarkable efficiency.
We also show that replacing the local one-bit mutation operator in the MAHH with the global bit-wise mutation operator, commonly used in EAs, yields a runtime of $min1, O(fraceln(n)m)m, O(nm)$ on JUMP functions.
arXiv Detail & Related papers (2024-07-19T12:10:05Z) - Agnostically Learning Multi-index Models with Queries [54.290489524576756]
We study the power of query access for the task of agnostic learning under the Gaussian distribution.
We show that query access gives significant runtime improvements over random examples for agnostically learning MIMs.
arXiv Detail & Related papers (2023-12-27T15:50:47Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Self-adjusting Population Sizes for the $(1, \lambda)$-EA on Monotone
Functions [7.111443975103329]
We study the $(1,lambda)$-EA with mutation rate $c/n$ for $cle 1$, where the population size is adaptively controlled with the $(1:s+1)$-success rule.
We show that this setup with $c=1$ is efficient on onemax for $s1$, but inefficient if $s ge 18$.
arXiv Detail & Related papers (2022-04-01T15:46:12Z) - TURF: A Two-factor, Universal, Robust, Fast Distribution Learning
Algorithm [64.13217062232874]
One of its most powerful and successful modalities approximates every distribution to an $ell$ distance essentially at most a constant times larger than its closest $t$-piece degree-$d_$.
We provide a method that estimates this number near-optimally, hence helps approach the best possible approximation.
arXiv Detail & Related papers (2022-02-15T03:49:28Z) - When Non-Elitism Meets Time-Linkage Problems [19.798298260257432]
We analyze on the influence of the non-elitism via comparing the performance of the elitist (1+$lambda$)EA and its non-elitist counterpart (1,$lambda$)EA.
We prove that with probability $1$, (1,$lambda$)EA can reach the global optimum and its expected runtime is $O(n3+clog n)$ with $lambda=c log_fracee-1 n$ for the constant $cge 1$.
arXiv Detail & Related papers (2021-04-14T13:03:42Z) - Private Stochastic Convex Optimization: Optimal Rates in $\ell_1$
Geometry [69.24618367447101]
Up to logarithmic factors the optimal excess population loss of any $(varepsilon,delta)$-differently private is $sqrtlog(d)/n + sqrtd/varepsilon n.$
We show that when the loss functions satisfy additional smoothness assumptions, the excess loss is upper bounded (up to logarithmic factors) by $sqrtlog(d)/n + (log(d)/varepsilon n)2/3.
arXiv Detail & Related papers (2021-03-02T06:53:44Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - Streaming Complexity of SVMs [110.63976030971106]
We study the space complexity of solving the bias-regularized SVM problem in the streaming model.
We show that for both problems, for dimensions of $frac1lambdaepsilon$, one can obtain streaming algorithms with spacely smaller than $frac1lambdaepsilon$.
arXiv Detail & Related papers (2020-07-07T17:10:00Z) - Optimal Mutation Rates for the $(1+\lambda)$ EA on OneMax [1.0965065178451106]
We extend the analysis of optimal mutation rates to two variants of the OneMax problem.
We compute for all population sizes $lambda in 2i mid 0 le i le 18$ which mutation rates minimize the expected running time.
Our results do not only provide a lower bound against which we can measure common evolutionary approaches.
arXiv Detail & Related papers (2020-06-20T01:23:14Z) - A Rigorous Runtime Analysis of the $(1 + (\lambda, \lambda))$ GA on Jump
Functions [8.34061303235504]
We conduct the first runtime analysis of this algorithm on a multimodal problem class, the jump functions benchmark.
For the isolated problem of leaving the local optimum of jump functions, we determine provably optimal parameters that lead to a runtime of $(n/k)k/2 eTheta(k)$.
arXiv Detail & Related papers (2020-04-14T17:54:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.