Random-key genetic algorithms: Principles and applications
- URL: http://arxiv.org/abs/2506.02120v2
- Date: Wed, 04 Jun 2025 17:44:05 GMT
- Title: Random-key genetic algorithms: Principles and applications
- Authors: Mariana A. Londe, Luciana S. Pessoa, Carlos E. Andrade, José F. Gonçalves, Mauricio G. C. Resende,
- Abstract summary: A random-key genetic algorithm is an evolutionary metaheuristic for discrete and global optimization.<n>This chapter reviews random-key genetic algorithms and describes an effective variant called biased random-key genetic algorithms.
- Score: 2.0971479389679337
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A random-key genetic algorithm is an evolutionary metaheuristic for discrete and global optimization. Each solution is encoded as a vector of N random keys, where a random key is a real number randomly generated in the continuous interval [0, 1). A decoder maps each vector of random keys to a solution of the optimization problem being solved and computes its cost. The benefit of this approach is that all genetic operators and transformations can be maintained within the unitary hypercube, regardless of the problem being addressed. This enhances the productivity and maintainability of the core framework. The algorithm starts with a population of P vectors of random keys. At each iteration, the vectors are partitioned into two sets: a smaller set of high-valued elite solutions and the remaining non-elite solutions. All elite elements are copied, without change, to the next population. A small number of random-key vectors (the mutants) is added to the population of the next iteration. The remaining elements of the population of the next iteration are generated by combining, with the parametrized uniform crossover of Spears and DeJong (1991), pairs of solutions. This chapter reviews random-key genetic algorithms and describes an effective variant called biased random-key genetic algorithms.
Related papers
- Optimizing Genetic Algorithms Using the Binomial Distribution [0.0]
Execution speed depends strongly on how we implement randomness.<n>We show how to optimize bit-flip mutation, uniform crossover, and control loop.<n>We implement our approach in the open-source Java library Chips-n-Salsa.
arXiv Detail & Related papers (2024-12-02T22:26:47Z) - A Novel Normalized-Cut Solver with Nearest Neighbor Hierarchical
Initialization [107.07093621337084]
Normalized-Cut (N-Cut) is a famous model of spectral clustering.
Traditional N-Cut solvers are two-stage: 1) calculating the continuous spectral embedding of normalized Laplacian matrix; 2) discretization via $K$-means or spectral rotation.
We propose a novel N-Cut solver based on the famous coordinate descent method.
arXiv Detail & Related papers (2023-11-26T07:11:58Z) - A Survey and Analysis of Evolutionary Operators for Permutations [0.0]
Evolving permutations directly requires specialized evolutionary operators.
In this paper, we survey the breadth of evolutionary operators for permutations.
We implement all of these in Chips-n-Salsa, an open source Java library for evolutionary computation.
arXiv Detail & Related papers (2023-11-24T16:32:44Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - A Hybrid Genetic Algorithm for the min-max Multiple Traveling Salesman
Problem [0.9790236766474201]
This paper proposes a hybrid genetic algorithm for solving the Multiple Traveling Salesman Problem (mTSP)
A novel crossover operator is designed to combine similar tours from two parents and offers great diversity for the population.
Our algorithm outperforms all existing algorithms on average, with similar cutoff time thresholds, when tested against benchmark sets.
arXiv Detail & Related papers (2023-07-14T01:57:10Z) - Population based change-point detection for the identification of
homozygosity islands [0.0]
We introduce a penalized maximum likelihood approach that can be efficiently computed by a dynamic programming algorithm or approximated by a fast greedy binary splitting algorithm.
We prove both algorithms converge almost surely to the set of change-points under very general assumptions on the distribution and independent sampling of the random vector.
This new approach is motivated by the problem of identifying homozygosity islands on the genome of individuals in a population.
arXiv Detail & Related papers (2021-11-19T12:53:41Z) - Sparse Quadratic Optimisation over the Stiefel Manifold with Application
to Permutation Synchronisation [71.27989298860481]
We address the non- optimisation problem of finding a matrix on the Stiefel manifold that maximises a quadratic objective function.
We propose a simple yet effective sparsity-promoting algorithm for finding the dominant eigenspace matrix.
arXiv Detail & Related papers (2021-09-30T19:17:35Z) - Adaptive Random Quantum Eigensolver [0.0]
We introduce a general method to parametrize and optimize the probability density function of a random number generator.
Our optimization is based on two figures of merit: learning speed and learning accuracy.
arXiv Detail & Related papers (2021-06-28T12:01:05Z) - Stochastic Saddle-Point Optimization for Wasserstein Barycenters [69.68068088508505]
We consider the populationimation barycenter problem for random probability measures supported on a finite set of points and generated by an online stream of data.
We employ the structure of the problem and obtain a convex-concave saddle-point reformulation of this problem.
In the setting when the distribution of random probability measures is discrete, we propose an optimization algorithm and estimate its complexity.
arXiv Detail & Related papers (2020-06-11T19:40:38Z) - Optimal Randomized First-Order Methods for Least-Squares Problems [56.05635751529922]
This class of algorithms encompasses several randomized methods among the fastest solvers for least-squares problems.
We focus on two classical embeddings, namely, Gaussian projections and subsampled Hadamard transforms.
Our resulting algorithm yields the best complexity known for solving least-squares problems with no condition number dependence.
arXiv Detail & Related papers (2020-02-21T17:45:32Z) - Optimal Iterative Sketching with the Subsampled Randomized Hadamard
Transform [64.90148466525754]
We study the performance of iterative sketching for least-squares problems.
We show that the convergence rate for Haar and randomized Hadamard matrices are identical, andally improve upon random projections.
These techniques may be applied to other algorithms that employ randomized dimension reduction.
arXiv Detail & Related papers (2020-02-03T16:17:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.