Genetic Programming is Naturally Suited to Evolve Bagging Ensembles
- URL: http://arxiv.org/abs/2009.06037v5
- Date: Fri, 5 Feb 2021 13:15:42 GMT
- Title: Genetic Programming is Naturally Suited to Evolve Bagging Ensembles
- Authors: Marco Virgolin
- Abstract summary: We show that minor changes to fitness evaluation and selection are sufficient to make a simple and otherwise-traditional GP algorithm evolve efficiently.
Our algorithm fares very well against state-of-the-art ensemble and non-ensemble GP algorithms.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning ensembles by bagging can substantially improve the generalization
performance of low-bias, high-variance estimators, including those evolved by
Genetic Programming (GP). To be efficient, modern GP algorithms for evolving
(bagging) ensembles typically rely on several (often inter-connected)
mechanisms and respective hyper-parameters, ultimately compromising ease of
use. In this paper, we provide experimental evidence that such complexity might
not be warranted. We show that minor changes to fitness evaluation and
selection are sufficient to make a simple and otherwise-traditional GP
algorithm evolve ensembles efficiently. The key to our proposal is to exploit
the way bagging works to compute, for each individual in the population,
multiple fitness values (instead of one) at a cost that is only marginally
higher than the one of a normal fitness evaluation. Experimental comparisons on
classification and regression tasks taken and reproduced from prior studies
show that our algorithm fares very well against state-of-the-art ensemble and
non-ensemble GP algorithms. We further provide insights into the proposed
approach by (i) scaling the ensemble size, (ii) ablating the changes to
selection, (iii) observing the evolvability induced by traditional subtree
variation. Code: https://github.com/marcovirgolin/2SEGP.
Related papers
- Improving Time and Memory Efficiency of Genetic Algorithms by Storing
Populations as Minimum Spanning Trees of Patches [0.0]
In evolutionary algorithms the computational cost of applying operators and storing populations is comparable to the cost of fitness evaluation.
We show that storing the population as a minimum spanning tree, where vertices correspond to individuals but only contain meta-information about them, is a viable alternative to the straightforward implementation.
Our experiments suggest that significant, even improvements -- including execution of crossover operators! -- can be achieved in terms of both memory usage and computational costs.
arXiv Detail & Related papers (2023-06-29T05:12:14Z) - The FAIRy Tale of Genetic Algorithms [1.0957528713294875]
We have extended Findable, Accessible, Interoperable and Reusable (FAIR) data principles to enable Genetic and reusability of algorithms.
We have presented an overview of methodological developments and variants of GA that makes it challenging to reproduce or even find the right source.
This work can be extended to numerous machine learning algorithms/methods.
arXiv Detail & Related papers (2023-04-29T11:36:09Z) - A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive
Coding Networks [65.34977803841007]
Predictive coding networks are neuroscience-inspired models with roots in both Bayesian statistics and neuroscience.
We show how by simply changing the temporal scheduling of the update rule for the synaptic weights leads to an algorithm that is much more efficient and stable than the original one.
arXiv Detail & Related papers (2022-11-16T00:11:04Z) - Coefficient Mutation in the Gene-pool Optimal Mixing Evolutionary
Algorithm for Symbolic Regression [0.0]
GP-GOMEA is a top-performing algorithm for symbolic regression.
We show how simple approaches for optimizing coefficients can be integrated into GP-GOMEA.
We find that coefficient mutation can help re-discover the underlying equation by a substantial amount.
arXiv Detail & Related papers (2022-04-26T08:58:47Z) - Scaling Gaussian Process Optimization by Evaluating a Few Unique
Candidates Multiple Times [119.41129787351092]
We show that sequential black-box optimization based on GPs can be made efficient by sticking to a candidate solution for multiple evaluation steps.
We modify two well-established GP-Opt algorithms, GP-UCB and GP-EI to adapt rules from batched GP-Opt.
arXiv Detail & Related papers (2022-01-30T20:42:14Z) - Improving RNA Secondary Structure Design using Deep Reinforcement
Learning [69.63971634605797]
We propose a new benchmark of applying reinforcement learning to RNA sequence design, in which the objective function is defined to be the free energy in the sequence's secondary structure.
We show results of the ablation analysis that we do for these algorithms, as well as graphs indicating the algorithm's performance across batches.
arXiv Detail & Related papers (2021-11-05T02:54:06Z) - Incremental Ensemble Gaussian Processes [53.3291389385672]
We propose an incremental ensemble (IE-) GP framework, where an EGP meta-learner employs an it ensemble of GP learners, each having a unique kernel belonging to a prescribed kernel dictionary.
With each GP expert leveraging the random feature-based approximation to perform online prediction and model update with it scalability, the EGP meta-learner capitalizes on data-adaptive weights to synthesize the per-expert predictions.
The novel IE-GP is generalized to accommodate time-varying functions by modeling structured dynamics at the EGP meta-learner and within each GP learner.
arXiv Detail & Related papers (2021-10-13T15:11:25Z) - Harnessing Heterogeneity: Learning from Decomposed Feedback in Bayesian
Modeling [68.69431580852535]
We introduce a novel GP regression to incorporate the subgroup feedback.
Our modified regression has provably lower variance -- and thus a more accurate posterior -- compared to previous approaches.
We execute our algorithm on two disparate social problems.
arXiv Detail & Related papers (2021-07-07T03:57:22Z) - AdaLead: A simple and robust adaptive greedy search algorithm for
sequence design [55.41644538483948]
We develop an easy-to-directed, scalable, and robust evolutionary greedy algorithm (AdaLead)
AdaLead is a remarkably strong benchmark that out-competes more complex state of the art approaches in a variety of biologically motivated sequence design challenges.
arXiv Detail & Related papers (2020-10-05T16:40:38Z) - Single- and Multi-Objective Evolutionary Algorithms for the Knapsack
Problem with Dynamically Changing Constraints [13.896724650508087]
We investigate single- and multi-objective baseline evolutionary algorithms for the classical knapsack problem.
Our results show that the multi-objective approaches using a population that caters for dynamic changes have a clear advantage on many benchmarks scenarios.
arXiv Detail & Related papers (2020-04-27T03:50:24Z) - On the Performance of Metaheuristics: A Different Perspective [0.0]
We study some basic evolutionary and swam-intelligence metaheuristics i.e. Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Artificial Bee Colony (ABC), Teaching-Learning-Based Optimization (TLBO) and Cuckoo Optimization algorithm (COA)
A large number of experiments have been conducted on 20 different optimization benchmark functions with different characteristics, and the results reveal to us some fundamental conclusions besides the following ranking order among these metaheuristics.
arXiv Detail & Related papers (2020-01-24T09:34:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.