Supervised Learning Achieves Human-Level Performance in MOBA Games: A
Case Study of Honor of Kings
- URL: http://arxiv.org/abs/2011.12582v1
- Date: Wed, 25 Nov 2020 08:45:55 GMT
- Title: Supervised Learning Achieves Human-Level Performance in MOBA Games: A
Case Study of Honor of Kings
- Authors: Deheng Ye, Guibin Chen, Peilin Zhao, Fuhao Qiu, Bo Yuan, Wen Zhang,
Sheng Chen, Mingfei Sun, Xiaoqian Li, Siqin Li, Jing Liang, Zhenjie Lian, Bei
Shi, Liang Wang, Tengfei Shi, Qiang Fu, Wei Yang, Lanxiao Huang
- Abstract summary: We present JueWu-SL, the first supervised-learning-based artificial intelligence (AI) program that achieves human-level performance in online battle arena (MOBA) games.
We integrate the macro-strategy and the micromanagement of MOBA-game-playing into neural networks in a supervised and end-to-end manner.
Tested on Honor of Kings, the most popular MOBA at present, our AI performs competitively at the level of High King players in standard 5v5 games.
- Score: 37.534249771219926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present JueWu-SL, the first supervised-learning-based artificial
intelligence (AI) program that achieves human-level performance in playing
multiplayer online battle arena (MOBA) games. Unlike prior attempts, we
integrate the macro-strategy and the micromanagement of MOBA-game-playing into
neural networks in a supervised and end-to-end manner. Tested on Honor of
Kings, the most popular MOBA at present, our AI performs competitively at the
level of High King players in standard 5v5 games.
Related papers
- DanZero+: Dominating the GuanDan Game through Reinforcement Learning [95.90682269990705]
We develop an AI program for an exceptionally complex and popular card game called GuanDan.
We first put forward an AI program named DanZero for this game.
In order to further enhance the AI's capabilities, we apply policy-based reinforcement learning algorithm to GuanDan.
arXiv Detail & Related papers (2023-12-05T08:07:32Z) - Mastering the Game of No-Press Diplomacy via Human-Regularized
Reinforcement Learning and Planning [95.78031053296513]
No-press Diplomacy is a complex strategy game involving both cooperation and competition.
We introduce a planning algorithm we call DiL-piKL that regularizes a reward-maximizing policy toward a human imitation-learned policy.
We show that DiL-piKL can be extended into a self-play reinforcement learning algorithm we call RL-DiL-piKL.
arXiv Detail & Related papers (2022-10-11T14:47:35Z) - Honor of Kings Arena: an Environment for Generalization in Competitive
Reinforcement Learning [36.43768953313382]
This paper introduces Honor of Kings Arena, a reinforcement learning environment based on Honor of Kings.
It presents new generalization challenges for competitive reinforcement learning.
arXiv Detail & Related papers (2022-09-18T06:29:27Z) - Learning Diverse Policies in MOBA Games via Macro-Goals [16.91587630049241]
We propose a novel Macro-Goals Guided framework, called MGG, to learn diverse policies in MOBA games.
MGG abstracts strategies as macro-goals from human demonstrations and trains a Meta-Controller to predict these macro-goals.
We show that MGG can execute diverse policies in different matches and lineups, and also outperform the state-of-the-art methods over 102 heroes.
arXiv Detail & Related papers (2021-10-27T07:15:42Z) - Towards Playing Full MOBA Games with Deep Reinforcement Learning [34.153341961273554]
MOBA games, e.g., Honor of Kings, League of Legends, and Dota 2, pose grand challenges to AI systems.
We propose a MOBA AI learning paradigm that methodologically enables playing full MOBA games with deep reinforcement learning.
arXiv Detail & Related papers (2020-11-25T12:52:33Z) - TotalBotWar: A New Pseudo Real-time Multi-action Game Challenge and
Competition for AI [62.997667081978825]
TotalBotWar is a new pseudo real-time multi-action challenge for game AI.
The game is based on the popular TotalWar games series where players manage an army to defeat the opponent's one.
arXiv Detail & Related papers (2020-09-18T09:13:56Z) - Suphx: Mastering Mahjong with Deep Reinforcement Learning [114.68233321904623]
We design an AI for Mahjong, named Suphx, based on deep reinforcement learning with some newly introduced techniques.
Suphx has demonstrated stronger performance than most top human players in terms of stable rank.
This is the first time that a computer program outperforms most top human players in Mahjong.
arXiv Detail & Related papers (2020-03-30T16:18:16Z) - Neural MMO v1.3: A Massively Multiagent Game Environment for Training
and Evaluating Neural Networks [48.5733173329785]
We present Neural MMO, a massively multiagent game environment inspired by MMOs.
We discuss our progress on two more general challenges in multiagent systems engineering for AI research: distributed infrastructure and game IO.
arXiv Detail & Related papers (2020-01-31T18:50:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.