(1+1) Genetic Programming With Functionally Complete Instruction Sets
Can Evolve Boolean Conjunctions and Disjunctions with Arbitrarily Small Error
- URL: http://arxiv.org/abs/2303.07455v1
- Date: Mon, 13 Mar 2023 20:24:21 GMT
- Title: (1+1) Genetic Programming With Functionally Complete Instruction Sets
Can Evolve Boolean Conjunctions and Disjunctions with Arbitrarily Small Error
- Authors: Benjamin Doerr, Andrei Lissovoi, Pietro S. Oliveto
- Abstract summary: GP systems can evolve a conjunction of $n$ variables if they are equipped with the minimal required components.
We show that a GP system can evolve the exact target function in $O(ell n log2 n)$ in expectation, where $ell geq is a limit on the size of any accepted tree.
- Score: 16.075339182916128
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently it has been proven that simple GP systems can efficiently evolve a
conjunction of $n$ variables if they are equipped with the minimal required
components. In this paper, we make a considerable step forward by analysing the
behaviour and performance of a GP system for evolving a Boolean conjunction or
disjunction of $n$ variables using a complete function set that allows the
expression of any Boolean function of up to $n$ variables. First we rigorously
prove that a GP system using the complete truth table to evaluate the program
quality, and equipped with both the AND and OR operators and positive literals,
evolves the exact target function in $O(\ell n \log^2 n)$ iterations in
expectation, where $\ell \geq n$ is a limit on the size of any accepted tree.
Additionally, we show that when a polynomial sample of possible inputs is used
to evaluate the solution quality, conjunctions or disjunctions with any
polynomially small generalisation error can be evolved with probability $1 -
O(\log^2(n)/n)$. The latter result also holds if GP uses AND, OR and positive
and negated literals, thus has the power to express any Boolean function of $n$
distinct variables. To prove our results we introduce a super-multiplicative
drift theorem that gives significantly stronger runtime bounds when the
expected progress is only slightly super-linear in the distance from the
optimum.
Related papers
- Quantum computational complexity of matrix functions [2.7488316163114823]
We study the computational complexity of two primitive problems.
We consider four functions -- monomials, Chebyshevs, the time evolution function, and the inverse function.
arXiv Detail & Related papers (2024-10-17T18:00:03Z) - Decomposing Hard SAT Instances with Metaheuristic Optimization [52.03315747221343]
We introduce the notion of decomposition hardness (d-hardness)
We show that the d-hardness expresses an estimate of the hardness of $C$ w.r.t.
arXiv Detail & Related papers (2023-12-16T12:44:36Z) - Exponential Separation between Quantum and Classical Ordered Binary
Decision Diagrams, Reordering Method and Hierarchies [68.93512627479197]
We study quantum Ordered Binary Decision Diagrams($OBDD$) model.
We prove lower bounds and upper bounds for OBDD with arbitrary order of input variables.
We extend hierarchy for read$k$-times Ordered Binary Decision Diagrams ($k$-OBDD$) of width.
arXiv Detail & Related papers (2022-04-22T12:37:56Z) - TURF: A Two-factor, Universal, Robust, Fast Distribution Learning
Algorithm [64.13217062232874]
One of its most powerful and successful modalities approximates every distribution to an $ell$ distance essentially at most a constant times larger than its closest $t$-piece degree-$d_$.
We provide a method that estimates this number near-optimally, hence helps approach the best possible approximation.
arXiv Detail & Related papers (2022-02-15T03:49:28Z) - Computing the Newton-step faster than Hessian accumulation [8.147652597876862]
We show that given the computational graph of the function, this bound can be reduced to $O(mtau3)$, where $tau, m$ are the width and size of a tree-decomposition of the graph.
The proposed algorithm generalizes nonlinear optimal-control methods based on LQR to general optimization problems and provides non-trivial gains in iteration-complexity even in cases where the Hessian is dense.
arXiv Detail & Related papers (2021-08-02T11:22:08Z) - Lockout: Sparse Regularization of Neural Networks [0.0]
Regularization is applied to improve accuracy by placing a constraint $P(w)leq t$ on the values of the parameters $w$.
We present a fast algorithm that provides all such solutions for any differentiable function $f$ and loss $L$, and any constraint $P$ that is an increasing monotone function of the absolute value of each parameter.
arXiv Detail & Related papers (2021-07-15T07:17:20Z) - Finding Global Minima via Kernel Approximations [90.42048080064849]
We consider the global minimization of smooth functions based solely on function evaluations.
In this paper, we consider an approach that jointly models the function to approximate and finds a global minimum.
arXiv Detail & Related papers (2020-12-22T12:59:30Z) - Towards Tight Communication Lower Bounds for Distributed Optimisation [30.134447658488057]
We consider a standard distributed optimisation setting where $N$ machines aim to jointly minimise the sum of the functions $sum_i = 1N f_i (x)$.
Our main result provides the first fully unconditional bounds on total number of bits which need to be sent and received by the $N$ machines.
We show that $Omega( Nd log d / Nvarepsilon)$ total bits need to be communicated between the machines to find an additive $epsilon$-approximation to the minimum of $sum_i = 1N
arXiv Detail & Related papers (2020-10-16T08:10:02Z) - An Optimal Separation of Randomized and Quantum Query Complexity [67.19751155411075]
We prove that for every decision tree, the absolute values of the Fourier coefficients of a given order $ellsqrtbinomdell (1+log n)ell-1,$ sum to at most $cellsqrtbinomdell (1+log n)ell-1,$ where $n$ is the number of variables, $d$ is the tree depth, and $c>0$ is an absolute constant.
arXiv Detail & Related papers (2020-08-24T06:50:57Z) - On the Modularity of Hypernetworks [103.1147622394852]
We show that for a structured target function, the overall number of trainable parameters in a hypernetwork is smaller by orders of magnitude than the number of trainable parameters of a standard neural network and an embedding method.
arXiv Detail & Related papers (2020-02-23T22:51:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.