Leveraging Transformers for StarCraft Macromanagement Prediction
- URL: http://arxiv.org/abs/2110.05343v1
- Date: Mon, 11 Oct 2021 15:12:21 GMT
- Title: Leveraging Transformers for StarCraft Macromanagement Prediction
- Authors: Muhammad Junaid Khan, Shah Hassan and Gita Sukthankar
- Abstract summary: We introduce a transformer-based neural architecture for two key StarCraft II macromanagement tasks: global state and build order prediction.
Unlike recurrent neural networks which suffer from a recency bias, transformers are able to capture patterns across very long time horizons.
One key advantage of transformers is their ability to generalize well, and we demonstrate that our model achieves an even better accuracy when used in a transfer learning setting.
- Score: 1.5469452301122177
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inspired by the recent success of transformers in natural language processing
and computer vision applications, we introduce a transformer-based neural
architecture for two key StarCraft II (SC2) macromanagement tasks: global state
and build order prediction. Unlike recurrent neural networks which suffer from
a recency bias, transformers are able to capture patterns across very long time
horizons, making them well suited for full game analysis. Our model utilizes
the MSC (Macromanagement in StarCraft II) dataset and improves on the top
performing gated recurrent unit (GRU) architecture in predicting global state
and build order as measured by mean accuracy over multiple time horizons. We
present ablation studies on our proposed architecture that support our design
decisions. One key advantage of transformers is their ability to generalize
well, and we demonstrate that our model achieves an even better accuracy when
used in a transfer learning setting in which models trained on games with one
racial matchup (e.g., Terran vs. Protoss) are transferred to a different one.
We believe that transformers' ability to model long games, potential for
parallelization, and generalization performance make them an excellent choice
for StarCraft agents.
Related papers
- Emergent Agentic Transformer from Chain of Hindsight Experience [96.56164427726203]
We show that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
This is the first time that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
arXiv Detail & Related papers (2023-05-26T00:43:02Z) - Transformer-based World Models Are Happy With 100k Interactions [0.4588028371034407]
We apply a transformer to real-world episodes in an autoregressive manner to build a sample-efficient world model.
The transformer allows our world model to access previous states directly, instead of viewing them through a compressed recurrent state.
By utilizing the Transformer-XL architecture, it is able to learn long-term dependencies while staying computationally efficient.
arXiv Detail & Related papers (2023-03-13T13:43:59Z) - Multi-Game Decision Transformers [49.257185338595434]
We show that a single transformer-based model can play a suite of up to 46 Atari games simultaneously at close-to-human performance.
We compare several approaches in this multi-game setting, such as online and offline RL methods and behavioral cloning.
We find that our Multi-Game Decision Transformer models offer the best scalability and performance.
arXiv Detail & Related papers (2022-05-30T16:55:38Z) - Improving Sample Efficiency of Value Based Models Using Attention and
Vision Transformers [52.30336730712544]
We introduce a deep reinforcement learning architecture whose purpose is to increase sample efficiency without sacrificing performance.
We propose a visually attentive model that uses transformers to learn a self-attention mechanism on the feature maps of the state representation.
We demonstrate empirically that this architecture improves sample complexity for several Atari environments, while also achieving better performance in some of the games.
arXiv Detail & Related papers (2022-02-01T19:03:03Z) - Vision Transformers for Dense Prediction [77.34726150561087]
We introduce dense vision transformers, an architecture that leverages vision transformers in place of convolutional networks as a backbone for dense prediction tasks.
Our experiments show that this architecture yields substantial improvements on dense prediction tasks.
arXiv Detail & Related papers (2021-03-24T18:01:17Z) - Transformers Solve the Limited Receptive Field for Monocular Depth
Prediction [82.90445525977904]
We propose TransDepth, an architecture which benefits from both convolutional neural networks and transformers.
This is the first paper which applies transformers into pixel-wise prediction problems involving continuous labels.
arXiv Detail & Related papers (2021-03-22T18:00:13Z) - Deep Policy Networks for NPC Behaviors that Adapt to Changing Design
Parameters in Roguelike Games [137.86426963572214]
Turn-based strategy games like Roguelikes, for example, present unique challenges to Deep Reinforcement Learning (DRL)
We propose two network architectures to better handle complex categorical state spaces and to mitigate the need for retraining forced by design decisions.
arXiv Detail & Related papers (2020-12-07T08:47:25Z) - Transformer Networks for Trajectory Forecasting [11.802437934289062]
We propose the novel use of Transformer Networks for trajectory forecasting.
This is a fundamental switch from the sequential step-by-step processing of LSTMs to the only-attention-based memory mechanisms of Transformers.
arXiv Detail & Related papers (2020-03-18T09:17:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.