Probabilistic Lexicase Selection
- URL: http://arxiv.org/abs/2305.11681v1
- Date: Fri, 19 May 2023 13:57:04 GMT
- Title: Probabilistic Lexicase Selection
- Authors: Li Ding, Edward Pantridge, Lee Spector
- Abstract summary: We introduce probabilistic lexicase selection (plexicase selection), a novel parent selection algorithm that efficiently approximates the probability distribution of lexicase selection.
Our method not only demonstrates superior problem-solving capabilities as a semantic-aware selection method, but also benefits from having a probabilistic representation of the selection process.
- Score: 6.177959045971966
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lexicase selection is a widely used parent selection algorithm in genetic
programming, known for its success in various task domains such as program
synthesis, symbolic regression, and machine learning. Due to its non-parametric
and recursive nature, calculating the probability of each individual being
selected by lexicase selection has been proven to be an NP-hard problem, which
discourages deeper theoretical understanding and practical improvements to the
algorithm. In this work, we introduce probabilistic lexicase selection
(plexicase selection), a novel parent selection algorithm that efficiently
approximates the probability distribution of lexicase selection. Our method not
only demonstrates superior problem-solving capabilities as a semantic-aware
selection method, but also benefits from having a probabilistic representation
of the selection process for enhanced efficiency and flexibility. Experiments
are conducted in two prevalent domains in genetic programming: program
synthesis and symbolic regression, using standard benchmarks including PSB and
SRBench. The empirical results show that plexicase selection achieves
state-of-the-art problem-solving performance that is competitive to the
lexicase selection, and significantly outperforms lexicase selection in
computation efficiency.
Related papers
- Switchable Decision: Dynamic Neural Generation Networks [98.61113699324429]
We propose a switchable decision to accelerate inference by dynamically assigning resources for each data instance.
Our method benefits from less cost during inference while keeping the same accuracy.
arXiv Detail & Related papers (2024-05-07T17:44:54Z) - DALex: Lexicase-like Selection via Diverse Aggregation [6.394522608656896]
We show that DALex (for Diversely Aggregated Lexicase) achieves significant speedups over lexicase selection and its relaxed variants.
Results on program synthesis, deep learning, symbolic regression, and learning systems demonstrate that DALex achieves significant speedups over lexicase selection and its relaxed variants.
arXiv Detail & Related papers (2024-01-23T01:20:15Z) - Provably Efficient UCB-type Algorithms For Learning Predictive State
Representations [55.00359893021461]
The sequential decision-making problem is statistically learnable if it admits a low-rank structure modeled by predictive state representations (PSRs)
This paper proposes the first known UCB-type approach for PSRs, featuring a novel bonus term that upper bounds the total variation distance between the estimated and true models.
In contrast to existing approaches for PSRs, our UCB-type algorithms enjoy computational tractability, last-iterate guaranteed near-optimal policy, and guaranteed model accuracy.
arXiv Detail & Related papers (2023-07-01T18:35:21Z) - Down-Sampled Epsilon-Lexicase Selection for Real-World Symbolic
Regression Problems [1.8047694351309207]
Down-sampled epsilon-lexicase selection combines epsilon-lexicase selection with random subsampling to improve the performance in the domain of symbolic regression.
We find that the diversity is reduced by using down-sampled epsilon-lexicase selection compared to standard epsilon-lexicase selection.
With down-sampled epsilon-lexicase selection we observe an improvement of the solution quality of up to 85% in comparison to standard epsilon-lexicase selection.
arXiv Detail & Related papers (2023-02-08T19:36:26Z) - Calculating lexicase selection probabilities is NP-Hard [0.0]
I prove that this problem, which I name lex-prob, is NP-Hard.
I achieve this proof by reducing SAT, a well-known NP-Complete problem, to lex-prob in time.
arXiv Detail & Related papers (2023-01-17T06:51:44Z) - Lexicase Selection at Scale [5.4968949435821735]
Lexicase selection is a semantic-aware parent selection method, which assesses individual test cases in a randomly-shuffled data stream.
One potential drawback of lexicase selection and its variants is that the selection procedure requires evaluating training cases in a single data stream.
We propose a novel method, fast lexicase selection, which incorporates lexicase selection and weighted shuffle with partial evaluation.
arXiv Detail & Related papers (2022-08-23T03:58:47Z) - Black-box Selective Inference via Bootstrapping [5.960626580825523]
Conditional selective inference requires an exact characterization of the selection event, which is often unavailable except for a few examples like the lasso.
This work addresses this challenge by introducing a generic approach to estimate the selection event, facilitating feasible inference conditioned on the selection event.
arXiv Detail & Related papers (2022-03-28T05:18:21Z) - Fast Feature Selection with Fairness Constraints [49.142308856826396]
We study the fundamental problem of selecting optimal features for model construction.
This problem is computationally challenging on large datasets, even with the use of greedy algorithm variants.
We extend the adaptive query model, recently proposed for the greedy forward selection for submodular functions, to the faster paradigm of Orthogonal Matching Pursuit for non-submodular functions.
The proposed algorithm achieves exponentially fast parallel run time in the adaptive query model, scaling much better than prior work.
arXiv Detail & Related papers (2022-02-28T12:26:47Z) - Online Active Model Selection for Pre-trained Classifiers [72.84853880948894]
We design an online selective sampling approach that actively selects informative examples to label and outputs the best model with high probability at any round.
Our algorithm can be used for online prediction tasks for both adversarial and streams.
arXiv Detail & Related papers (2020-10-19T19:53:15Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - Extreme Algorithm Selection With Dyadic Feature Representation [78.13985819417974]
We propose the setting of extreme algorithm selection (XAS) where we consider fixed sets of thousands of candidate algorithms.
We assess the applicability of state-of-the-art AS techniques to the XAS setting and propose approaches leveraging a dyadic feature representation.
arXiv Detail & Related papers (2020-01-29T09:40:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.