Portfolio Search and Optimization for General Strategy Game-Playing
- URL: http://arxiv.org/abs/2104.10429v1
- Date: Wed, 21 Apr 2021 09:28:28 GMT
- Title: Portfolio Search and Optimization for General Strategy Game-Playing
- Authors: Alexander Dockhorn, Jorge Hurtado-Grueso, Dominik Jeurissen, Linjie
Xu, Diego Perez-Liebana
- Abstract summary: We propose a new algorithm for optimization and action-selection based on the Rolling Horizon Evolutionary Algorithm.
For the optimization of the agents' parameters and portfolio sets we study the use of the N-tuple Bandit Evolutionary Algorithm.
An analysis of the agents' performance shows that the proposed algorithm generalizes well to all game-modes and is able to outperform other portfolio methods.
- Score: 58.896302717975445
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Portfolio methods represent a simple but efficient type of action abstraction
which has shown to improve the performance of search-based agents in a range of
strategy games. We first review existing portfolio techniques and propose a new
algorithm for optimization and action-selection based on the Rolling Horizon
Evolutionary Algorithm. Moreover, a series of variants are developed to solve
problems in different aspects. We further analyze the performance of discussed
agents in a general strategy game-playing task. For this purpose, we run
experiments on three different game-modes of the Stratega framework. For the
optimization of the agents' parameters and portfolio sets we study the use of
the N-tuple Bandit Evolutionary Algorithm. The resulting portfolio sets suggest
a high diversity in play-styles while being able to consistently beat the
sample agents. An analysis of the agents' performance shows that the proposed
algorithm generalizes well to all game-modes and is able to outperform other
portfolio methods.
Related papers
- Deep Reinforcement Learning for Online Optimal Execution Strategies [49.1574468325115]
This paper tackles the challenge of learning non-Markovian optimal execution strategies in dynamic financial markets.
We introduce a novel actor-critic algorithm based on Deep Deterministic Policy Gradient (DDPG)
We show that our algorithm successfully approximates the optimal execution strategy.
arXiv Detail & Related papers (2024-10-17T12:38:08Z) - An accelerate Prediction Strategy for Dynamic Multi-Objective Optimization [7.272641346606365]
We introduce novel approaches for accelerating prediction strategies within the evolutionary algorithm framework.
We propose an adaptive prediction strategy that incorporates second-order derivatives to predict and adjust the algorithms search behavior.
We evaluate the performance of the proposed method against four state-of-the-art algorithms using standard DMOPs benchmark problems.
arXiv Detail & Related papers (2024-10-08T08:13:49Z) - Deep Reinforcement Learning for Dynamic Algorithm Selection: A
Proof-of-Principle Study on Differential Evolution [27.607740475924448]
We propose a deep reinforcement learning-based dynamic algorithm selection framework to accomplish this task.
We employ a sophisticated deep neural network model to infer the optimal action, ensuring informed algorithm selections.
As a proof-of-principle study, we apply this framework to a group of Differential Evolution algorithms.
arXiv Detail & Related papers (2024-03-04T15:40:28Z) - PS-AAS: Portfolio Selection for Automated Algorithm Selection in
Black-Box Optimization [4.842307002343618]
The performance of automated algorithm selection depends on the portfolio of algorithms to choose from.
In practice, probably the most common way to choose the algorithms for the portfolio is a greedy selection of the algorithms that perform well in some reference tasks of interest.
Our proposed method creates algorithm behavior meta-representations, constructs a graph from a set of algorithms based on their meta-representation similarity, and applies a graph algorithm to select a final portfolio of diverse, representative, and non-redundant algorithms.
arXiv Detail & Related papers (2023-10-14T12:13:41Z) - Local Optimization Achieves Global Optimality in Multi-Agent
Reinforcement Learning [139.53668999720605]
We present a multi-agent PPO algorithm in which the local policy of each agent is updated similarly to vanilla PPO.
We prove that with standard regularity conditions on the Markov game and problem-dependent quantities, our algorithm converges to the globally optimal policy at a sublinear rate.
arXiv Detail & Related papers (2023-05-08T16:20:03Z) - Backpropagation of Unrolled Solvers with Folded Optimization [55.04219793298687]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
One typical strategy is algorithm unrolling, which relies on automatic differentiation through the operations of an iterative solver.
This paper provides theoretical insights into the backward pass of unrolled optimization, leading to a system for generating efficiently solvable analytical models of backpropagation.
arXiv Detail & Related papers (2023-01-28T01:50:42Z) - A novel multiobjective evolutionary algorithm based on decomposition and
multi-reference points strategy [14.102326122777475]
Multiobjective evolutionary algorithm based on decomposition (MOEA/D) has been regarded as a significantly promising approach for solving multiobjective optimization problems (MOPs)
We propose an improved MOEA/D algorithm by virtue of the well-known Pascoletti-Serafini scalarization method and a new strategy of multi-reference points.
arXiv Detail & Related papers (2021-10-27T02:07:08Z) - Meta Navigator: Search for a Good Adaptation Policy for Few-shot
Learning [113.05118113697111]
Few-shot learning aims to adapt knowledge learned from previous tasks to novel tasks with only a limited amount of labeled data.
Research literature on few-shot learning exhibits great diversity, while different algorithms often excel at different few-shot learning scenarios.
We present Meta Navigator, a framework that attempts to solve the limitation in few-shot learning by seeking a higher-level strategy.
arXiv Detail & Related papers (2021-09-13T07:20:01Z) - PAMELI: A Meta-Algorithm for Computationally Expensive Multi-Objective
Optimization Problems [0.0]
The proposed algorithm is based on solving a set of surrogate problems defined by models of the real one.
Our algorithm also performs a meta-search for optimal surrogate models and navigation strategies for the optimization landscape.
arXiv Detail & Related papers (2021-03-19T11:18:03Z) - Generalization in portfolio-based algorithm selection [97.74604695303285]
We provide the first provable guarantees for portfolio-based algorithm selection.
We show that if the portfolio is large, overfitting is inevitable, even with an extremely simple algorithm selector.
arXiv Detail & Related papers (2020-12-24T16:33:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.