Early years of Biased Random-Key Genetic Algorithms: A systematic review
- URL: http://arxiv.org/abs/2405.01765v3
- Date: Thu, 23 May 2024 13:47:57 GMT
- Title: Early years of Biased Random-Key Genetic Algorithms: A systematic review
- Authors: Mariana A. Londe, Luciana S. Pessoa, Cartlos E. Andrade, Mauricio G. C. Resende,
- Abstract summary: This paper presents a systematic literature review and bibliometric analysis focusing on Biased Random-Key Genetic Algorithms (BRKGA)
BRKGA is a metaheuristic framework that uses random-key-based chromosomes with biased, uniform, and elitist mating strategies alongside a genetic algorithm.
- Score: 2.249916681499244
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a systematic literature review and bibliometric analysis focusing on Biased Random-Key Genetic Algorithms (BRKGA). BRKGA is a metaheuristic framework that uses random-key-based chromosomes with biased, uniform, and elitist mating strategies alongside a genetic algorithm. This review encompasses around~250 papers, covering a diverse array of applications ranging from classical combinatorial optimization problems to real-world industrial scenarios, and even non-traditional applications like hyperparameter tuning in machine learning and scenario generation for two-stage problems. In summary, this study offers a comprehensive examination of the BRKGA metaheuristic and its various applications, shedding light on key areas for future research.
Related papers
- A Learning Search Algorithm for the Restricted Longest Common Subsequence Problem [40.64116457007417]
The Restricted Longest Common Subsequence (RLCS) problem has significant applications in bioinformatics.
This paper introduces two novel approaches designed to enhance the search process by steering it towards promising regions.
An important contribution of this paper is found in the generation of real-world instances where scientific abstracts serve as input strings.
arXiv Detail & Related papers (2024-10-15T20:02:15Z) - Parallel Strategies for Best-First Generalized Planning [51.713634067802104]
Generalized planning (GP) is a research area of AI that studies the automated synthesis of algorithmic-like solutions capable of solving multiple classical planning instances.
One of the current advancements has been the introduction of Best-First Generalized Planning (BFGP), a GP algorithm based on a novel solution space that can be explored with search.
This paper evaluates the application of parallel search techniques to BFGP, another critical component in closing the performance gap.
arXiv Detail & Related papers (2024-07-31T09:50:22Z) - DNA Sequence Classification with Compressors [0.0]
Our study introduces a novel adaptation of Jiang et al.'s compressor-based, parameter-free classification method, specifically tailored for DNA sequence analysis.
Not only does this method align with the current state-of-the-art in terms of accuracy, but it also offers a more resource-efficient alternative to traditional machine learning methods.
arXiv Detail & Related papers (2024-01-25T09:17:19Z) - Biased Random-Key Genetic Algorithms: A Review [2.4578723416255754]
The review encompasses over 150 papers with a wide range of applications.
Scheduling is by far the most prevalent application area in this review, followed by network design and location problems.
The most frequent hybridization method employed is local search, and new features aim to increase population diversity.
arXiv Detail & Related papers (2023-12-01T22:32:58Z) - Regularization-Based Methods for Ordinal Quantification [49.606912965922504]
We study the ordinal case, i.e., the case in which a total order is defined on the set of n>2 classes.
We propose a novel class of regularized OQ algorithms, which outperforms existing algorithms in our experiments.
arXiv Detail & Related papers (2023-10-13T16:04:06Z) - Some Experiences with Hybrid Genetic Algorithms in Solving the
Uncapacitated Examination Timetabling Problem [0.0]
The paper provides experimental experiences on two local search hybridized genetic algorithms in solving the uncapacitated examination timetabling problem.
The proposed two hybrid algorithms use partition and priority based solution representations which are inspired from successful genetic algorithms proposed for graph coloring and project scheduling problems, respectively.
arXiv Detail & Related papers (2023-06-01T10:41:32Z) - Stochastic Gradient Descent-Ascent: Unified Theory and New Efficient
Methods [73.35353358543507]
Gradient Descent-Ascent (SGDA) is one of the most prominent algorithms for solving min-max optimization and variational inequalities problems (VIP)
In this paper, we propose a unified convergence analysis that covers a large variety of descent-ascent methods.
We develop several new variants of SGDA such as a new variance-reduced method (L-SVRGDA), new distributed methods with compression (QSGDA, DIANA-SGDA, VR-DIANA-SGDA), and a new method with coordinate randomization (SEGA-SGDA)
arXiv Detail & Related papers (2022-02-15T09:17:39Z) - Result Diversification by Multi-objective Evolutionary Algorithms with
Theoretical Guarantees [94.72461292387146]
We propose to reformulate the result diversification problem as a bi-objective search problem, and solve it by a multi-objective evolutionary algorithm (EA)
We theoretically prove that the GSEMO can achieve the optimal-time approximation ratio, $1/2$.
When the objective function changes dynamically, the GSEMO can maintain this approximation ratio in running time, addressing the open question proposed by Borodin et al.
arXiv Detail & Related papers (2021-10-18T14:00:22Z) - Hybrid Random Features [60.116392415715275]
We propose a new class of random feature methods for linearizing softmax and Gaussian kernels called hybrid random features (HRFs)
HRFs automatically adapt the quality of kernel estimation to provide most accurate approximation in the defined regions of interest.
arXiv Detail & Related papers (2021-10-08T20:22:59Z) - Fractal Structure and Generalization Properties of Stochastic
Optimization Algorithms [71.62575565990502]
We prove that the generalization error of an optimization algorithm can be bounded on the complexity' of the fractal structure that underlies its generalization measure.
We further specialize our results to specific problems (e.g., linear/logistic regression, one hidden/layered neural networks) and algorithms.
arXiv Detail & Related papers (2021-06-09T08:05:36Z) - Random Features for Kernel Approximation: A Survey on Algorithms,
Theory, and Beyond [35.32894170512829]
In this survey, we systematically review the work on random features from the past ten years.
First, the motivations, characteristics and contributions of representative random features based algorithms are summarized.
Second, we review theoretical results that center around the following key question: how many random features are needed to ensure a high approximation quality.
Third, we provide a comprehensive evaluation of popular random features based algorithms on several large-scale benchmark datasets.
arXiv Detail & Related papers (2020-04-23T13:44:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.