Evolutionary Multi-Armed Bandits with Genetic Thompson Sampling
- URL: http://arxiv.org/abs/2205.10113v1
- Date: Tue, 26 Apr 2022 22:41:17 GMT
- Title: Evolutionary Multi-Armed Bandits with Genetic Thompson Sampling
- Authors: Baihan Lin
- Abstract summary: We propose the Genetic Thompson Sampling, a bandit algorithm that keeps a population of agents and update them with genetic principles such as elite selection, crossover and mutations.
We also introduce EvoBandit, a web-based interactive visualization to guide the readers through the entire learning process and perform lightweight evaluations on the fly.
- Score: 13.173307471333619
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As two popular schools of machine learning, online learning and evolutionary
computations have become two important driving forces behind real-world
decision making engines for applications in biomedicine, economics, and
engineering fields. Although there are prior work that utilizes bandits to
improve evolutionary algorithms' optimization process, it remains a field of
blank on how evolutionary approach can help improve the sequential decision
making tasks of online learning agents such as the multi-armed bandits. In this
work, we propose the Genetic Thompson Sampling, a bandit algorithm that keeps a
population of agents and update them with genetic principles such as elite
selection, crossover and mutations. Empirical results in multi-armed bandit
simulation environments and a practical epidemic control problem suggest that
by incorporating the genetic algorithm into the bandit algorithm, our method
significantly outperforms the baselines in nonstationary settings. Lastly, we
introduce EvoBandit, a web-based interactive visualization to guide the readers
through the entire learning process and perform lightweight evaluations on the
fly. We hope to engage researchers into this growing field of research with
this investigation.
Related papers
- DARLEI: Deep Accelerated Reinforcement Learning with Evolutionary
Intelligence [77.78795329701367]
We present DARLEI, a framework that combines evolutionary algorithms with parallelized reinforcement learning.
We characterize DARLEI's performance under various conditions, revealing factors impacting diversity of evolved morphologies.
We hope to extend DARLEI in future work to include interactions between diverse morphologies in richer environments.
arXiv Detail & Related papers (2023-12-08T16:51:10Z) - Genetic Algorithm enhanced by Deep Reinforcement Learning in parent
selection mechanism and mutation : Minimizing makespan in permutation flow
shop scheduling problems [0.18846515534317265]
The proposed RL+GA method was specifically tested on the flow shop scheduling problem (FSP)
The hybrid algorithm incorporates neural networks (NN) and uses the off-policy method Q-learning.
Results of the study highlight the effectiveness of the RL+GA approach in improving the performance of the primitive GA.
arXiv Detail & Related papers (2023-11-10T08:51:42Z) - Supervised Pretraining Can Learn In-Context Reinforcement Learning [96.62869749926415]
In this paper, we study the in-context learning capabilities of transformers in decision-making problems.
We introduce and study Decision-Pretrained Transformer (DPT), a supervised pretraining method where the transformer predicts an optimal action.
We find that the pretrained transformer can be used to solve a range of RL problems in-context, exhibiting both exploration online and conservatism offline.
arXiv Detail & Related papers (2023-06-26T17:58:50Z) - A Reinforcement Learning-assisted Genetic Programming Algorithm for Team
Formation Problem Considering Person-Job Matching [70.28786574064694]
A reinforcement learning-assisted genetic programming algorithm (RL-GP) is proposed to enhance the quality of solutions.
The hyper-heuristic rules obtained through efficient learning can be utilized as decision-making aids when forming project teams.
arXiv Detail & Related papers (2023-04-08T14:32:12Z) - Generative Adversarial Neuroevolution for Control Behaviour Imitation [3.04585143845864]
We propose to explore whether deep neuroevolution can be used for behaviour imitation on popular simulation environments.
We introduce a simple co-evolutionary adversarial generation framework, and evaluate its capabilities by evolving standard deep recurrent networks.
Across all tasks, we find the final elite actor agents capable of achieving scores as high as those obtained by the pre-trained agents.
arXiv Detail & Related papers (2023-04-03T16:33:22Z) - Training and Evaluation of Deep Policies using Reinforcement Learning
and Generative Models [67.78935378952146]
GenRL is a framework for solving sequential decision-making problems.
It exploits the combination of reinforcement learning and latent variable generative models.
We experimentally determine the characteristics of generative models that have most influence on the performance of the final policy training.
arXiv Detail & Related papers (2022-04-18T22:02:32Z) - Direct Mutation and Crossover in Genetic Algorithms Applied to
Reinforcement Learning Tasks [0.9137554315375919]
This paper will focus on applying neuroevolution using a simple genetic algorithm (GA) to find the weights of a neural network that produce optimally behaving agents.
We present two novel modifications that improve the data efficiency and speed of convergence when compared to the initial implementation.
arXiv Detail & Related papers (2022-01-13T07:19:28Z) - Improving RNA Secondary Structure Design using Deep Reinforcement
Learning [69.63971634605797]
We propose a new benchmark of applying reinforcement learning to RNA sequence design, in which the objective function is defined to be the free energy in the sequence's secondary structure.
We show results of the ablation analysis that we do for these algorithms, as well as graphs indicating the algorithm's performance across batches.
arXiv Detail & Related papers (2021-11-05T02:54:06Z) - An Efficient Application of Neuroevolution for Competitive Multiagent
Learning [0.0]
NEAT is a popular evolutionary strategy used to obtain the best performing neural network architecture.
This paper utilizes the NEAT algorithm to achieve competitive multiagent learning on a modified pong game environment.
arXiv Detail & Related papers (2021-05-23T10:34:48Z) - Lineage Evolution Reinforcement Learning [15.469857142001482]
Lineage evolution reinforcement learning is a derivative algorithm which accords with the general agent population learning system.
Our experiments show that the idea of evolution with lineage improves the performance of original reinforcement learning algorithm in some games in Atari 2600.
arXiv Detail & Related papers (2020-09-26T11:58:16Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.