Insights From the NeurIPS 2021 NetHack Challenge
- URL: http://arxiv.org/abs/2203.11889v1
- Date: Tue, 22 Mar 2022 17:01:07 GMT
- Title: Insights From the NeurIPS 2021 NetHack Challenge
- Authors: Eric Hambro, Sharada Mohanty, Dmitrii Babaev, Minwoo Byeon, Dipam
Chakraborty, Edward Grefenstette, Minqi Jiang, Daejin Jo, Anssi Kanervisto,
Jongmin Kim, Sungwoong Kim, Robert Kirk, Vitaly Kurin, Heinrich K\"uttler,
Taehwon Kwon, Donghoon Lee, Vegard Mella, Nantas Nardelli, Ivan Nazarov,
Nikita Ovsov, Jack Parker-Holder, Roberta Raileanu, Karolis Ramanauskas, Tim
Rockt\"aschel, Danielle Rothermel, Mikayel Samvelyan, Dmitry Sorokin, Maciej
Sypetkowski, Micha{\l} Sypetkowski
- Abstract summary: The first NeurIPS 2021 NetHack Challenge showcased community-driven progress in AI.
It served as a direct comparison between neural (e.g., deep RL) and symbolic AI, as well as hybrid systems.
No agent got close to winning the game, illustrating NetHack's suitability as a long-term benchmark for AI research.
- Score: 40.52602443114554
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this report, we summarize the takeaways from the first NeurIPS 2021
NetHack Challenge. Participants were tasked with developing a program or agent
that can win (i.e., 'ascend' in) the popular dungeon-crawler game of NetHack by
interacting with the NetHack Learning Environment (NLE), a scalable,
procedurally generated, and challenging Gym environment for reinforcement
learning (RL). The challenge showcased community-driven progress in AI with
many diverse approaches significantly beating the previously best results on
NetHack. Furthermore, it served as a direct comparison between neural (e.g.,
deep RL) and symbolic AI, as well as hybrid systems, demonstrating that on
NetHack symbolic bots currently outperform deep RL by a large margin. Lastly,
no agent got close to winning the game, illustrating NetHack's suitability as a
long-term benchmark for AI research.
Related papers
- Playing NetHack with LLMs: Potential & Limitations as Zero-Shot Agents [0.0]
Large Language Models (LLMs) have shown great success as high-level planners for zero-shot game-playing agents.
We present NetPlay, the first LLM-powered zero-shot agent for the challenging roguelike NetHack.
arXiv Detail & Related papers (2024-03-01T17:22:16Z) - DanZero+: Dominating the GuanDan Game through Reinforcement Learning [95.90682269990705]
We develop an AI program for an exceptionally complex and popular card game called GuanDan.
We first put forward an AI program named DanZero for this game.
In order to further enhance the AI's capabilities, we apply policy-based reinforcement learning algorithm to GuanDan.
arXiv Detail & Related papers (2023-12-05T08:07:32Z) - LuckyMera: a Modular AI Framework for Building Hybrid NetHack Agents [7.23273667916516]
Roguelike video games offer a good trade-off in terms of complexity of the environment and computational costs.
We present LuckyMera, a flexible, modular, generalization and AI framework built around NetHack.
LuckyMera comes with a set of off-the-shelf symbolic and neural modules (called "skills"): these modules can be either hard-coded behaviors, or neural Reinforcement Learning approaches.
arXiv Detail & Related papers (2023-07-17T14:46:59Z) - NetHack is Hard to Hack [37.24009814390211]
In the NeurIPS 2021 NetHack Challenge, symbolic agents outperformed neural approaches by over four times in median game score.
We present an extensive study on neural policy learning for NetHack.
We produce a state-of-the-art neural agent that surpasses previous fully neural policies by 127% in offline settings and 25% in online settings on median game score.
arXiv Detail & Related papers (2023-05-30T17:30:17Z) - Dungeons and Data: A Large-Scale NetHack Dataset [19.5560914918284]
We present the NetHack Learning dataset (NLD), a large and highly-scalable dataset of trajectories from the popular game of NetHack.
NLD consists of three parts: 10 billion state transitions from 1.5 million human trajectories collected on the NAO public NetHack server from 2009 to 2020; 3 billion state-action-score transitions from 100,000 trajectories collected from the symbolic bot winner of the NetHack Challenge 2021.
We evaluate a wide range of existing algorithms including online and offline RL, as well as learning from demonstrations, showing that significant research advances are needed to fully leverage large-scale datasets for challenging sequential
arXiv Detail & Related papers (2022-11-01T15:43:29Z) - Mastering the Game of No-Press Diplomacy via Human-Regularized
Reinforcement Learning and Planning [95.78031053296513]
No-press Diplomacy is a complex strategy game involving both cooperation and competition.
We introduce a planning algorithm we call DiL-piKL that regularizes a reward-maximizing policy toward a human imitation-learned policy.
We show that DiL-piKL can be extended into a self-play reinforcement learning algorithm we call RL-DiL-piKL.
arXiv Detail & Related papers (2022-10-11T14:47:35Z) - DeepCrawl: Deep Reinforcement Learning for Turn-based Strategy Games [137.86426963572214]
We introduce DeepCrawl, a fully-playable Roguelike prototype for iOS and Android in which all agents are controlled by policy networks trained using Deep Reinforcement Learning (DRL)
Our aim is to understand whether recent advances in DRL can be used to develop convincing behavioral models for non-player characters in videogames.
arXiv Detail & Related papers (2020-12-03T13:53:29Z) - The NetHack Learning Environment [79.06395964379107]
We present the NetHack Learning Environment (NLE), a procedurally generated rogue-like environment for Reinforcement Learning research.
We argue that NetHack is sufficiently complex to drive long-term research on problems such as exploration, planning, skill acquisition, and language-conditioned RL.
We demonstrate empirical success for early stages of the game using a distributed Deep RL baseline and Random Network Distillation exploration.
arXiv Detail & Related papers (2020-06-24T14:12:56Z) - Suphx: Mastering Mahjong with Deep Reinforcement Learning [114.68233321904623]
We design an AI for Mahjong, named Suphx, based on deep reinforcement learning with some newly introduced techniques.
Suphx has demonstrated stronger performance than most top human players in terms of stable rank.
This is the first time that a computer program outperforms most top human players in Mahjong.
arXiv Detail & Related papers (2020-03-30T16:18:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.