The MineRL 2020 Competition on Sample Efficient Reinforcement Learning
using Human Priors
- URL: http://arxiv.org/abs/2101.11071v1
- Date: Tue, 26 Jan 2021 20:32:30 GMT
- Title: The MineRL 2020 Competition on Sample Efficient Reinforcement Learning
using Human Priors
- Authors: William H. Guss, Mario Ynocente Castro, Sam Devlin, Brandon Houghton,
Noboru Sean Kuno, Crissman Loomis, Stephanie Milani, Sharada Mohanty, Keisuke
Nakata, Ruslan Salakhutdinov, John Schulman, Shinya Shiroshita, Nicholay
Topin, Avinash Ummadisingu, Oriol Vinyals
- Abstract summary: We propose a second iteration of the MineRL Competition.
The primary goal of the competition is to foster the development of algorithms which can efficiently leverage human demonstrations.
The competition is structured into two rounds in which competitors are provided several paired versions of the dataset and environment.
At the end of each round, competitors submit containerized versions of their learning algorithms to the AIcrowd platform.
- Score: 62.9301667732188
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although deep reinforcement learning has led to breakthroughs in many
difficult domains, these successes have required an ever-increasing number of
samples, affording only a shrinking segment of the AI community access to their
development. Resolution of these limitations requires new, sample-efficient
methods. To facilitate research in this direction, we propose this second
iteration of the MineRL Competition. The primary goal of the competition is to
foster the development of algorithms which can efficiently leverage human
demonstrations to drastically reduce the number of samples needed to solve
complex, hierarchical, and sparse environments. To that end, participants
compete under a limited environment sample-complexity budget to develop systems
which solve the MineRL ObtainDiamond task in Minecraft, a sequential decision
making environment requiring long-term planning, hierarchical control, and
efficient exploration methods. The competition is structured into two rounds in
which competitors are provided several paired versions of the dataset and
environment with different game textures and shaders. At the end of each round,
competitors submit containerized versions of their learning algorithms to the
AIcrowd platform where they are trained from scratch on a hold-out
dataset-environment pair for a total of 4-days on a pre-specified hardware
platform. In this follow-up iteration to the NeurIPS 2019 MineRL Competition,
we implement new features to expand the scale and reach of the competition. In
response to the feedback of the previous participants, we introduce a second
minor track focusing on solutions without access to environment interactions of
any kind except during test-time. Further we aim to prompt domain agnostic
submissions by implementing several novel competition mechanics including
action-space randomization and desemantization of observations and actions.
Related papers
- Benchmarking Robustness and Generalization in Multi-Agent Systems: A
Case Study on Neural MMO [50.58083807719749]
We present the results of the second Neural MMO challenge, hosted at IJCAI 2022, which received 1600+ submissions.
This competition targets robustness and generalization in multi-agent systems.
We will open-source our benchmark including the environment wrapper, baselines, a visualization tool, and selected policies for further research.
arXiv Detail & Related papers (2023-08-30T07:16:11Z) - Tackling Diverse Minorities in Imbalanced Classification [80.78227787608714]
Imbalanced datasets are commonly observed in various real-world applications, presenting significant challenges in training classifiers.
We propose generating synthetic samples iteratively by mixing data samples from both minority and majority classes.
We demonstrate the effectiveness of our proposed framework through extensive experiments conducted on seven publicly available benchmark datasets.
arXiv Detail & Related papers (2023-08-28T18:48:34Z) - DIAMBRA Arena: a New Reinforcement Learning Platform for Research and
Experimentation [91.3755431537592]
This work presents DIAMBRA Arena, a new platform for reinforcement learning research and experimentation.
It features a collection of high-quality environments exposing a Python API fully compliant with OpenAI Gym standard.
They are episodic tasks with discrete actions and observations composed by raw pixels plus additional numerical values.
arXiv Detail & Related papers (2022-10-19T14:39:10Z) - Lessons learned from the NeurIPS 2021 MetaDL challenge: Backbone
fine-tuning without episodic meta-learning dominates for few-shot learning
image classification [40.901760230639496]
We describe the design of the MetaDL competition series, the datasets, the best experimental results, and the top-ranked methods in the NeurIPS 2021 challenge.
The solutions of the top participants have been open-sourced.
arXiv Detail & Related papers (2022-06-15T10:27:23Z) - Retrospective on the 2021 BASALT Competition on Learning from Human
Feedback [92.37243979045817]
The goal of the competition was to promote research towards agents that use learning from human feedback (LfHF) techniques to solve open-world tasks.
Rather than mandating the use of LfHF techniques, we described four tasks in natural language to be accomplished in the video game Minecraft.
Teams developed a diverse range of LfHF algorithms across a variety of possible human feedback types.
arXiv Detail & Related papers (2022-04-14T17:24:54Z) - MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned [60.11039031794829]
Reinforcement learning competitions advance the field by providing appropriate scope and support to develop solutions toward a specific problem.
We hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.
The participants of this easier track were able to obtain a diamond, and the participants of the harder track progressed the generalizable solutions in the same task.
arXiv Detail & Related papers (2022-02-17T13:37:35Z) - SEIHAI: A Sample-efficient Hierarchical AI for the MineRL Competition [32.635756704572266]
We present textbfSEIHAI, a textbfSample-textbfefftextbficient textbfHierarchical textbfAI, that takes advantage of the human demonstrations and the task structure.
Specifically, we split the task into several sequentially dependent subtasks, and train a suitable agent for each subtask using reinforcement learning and imitation learning.
SEIHAI takes the first place in the preliminary and final of the NeurIPS-2020 MineRL competition.
arXiv Detail & Related papers (2021-11-17T01:36:40Z) - Towards robust and domain agnostic reinforcement learning competitions [12.731614722371376]
Reinforcement learning competitions have formed the basis for standard research benchmarks.
Despite this, a majority of challenges suffer from the same fundamental problems.
We present a new framework of competition design that promotes the development of algorithms that overcome these barriers.
arXiv Detail & Related papers (2021-06-07T16:15:46Z) - Retrospective Analysis of the 2019 MineRL Competition on Sample
Efficient Reinforcement Learning [27.440055101691115]
We held the MineRL Competition on Sample Efficient Reinforcement Learning Using Human Priors at the Thirty-third Conference on Neural Information Processing Systems (NeurIPS)
The primary goal of this competition was to promote the development of algorithms that use human demonstrations alongside reinforcement learning to reduce the number of samples needed to solve complex, hierarchical, and sparse environments.
arXiv Detail & Related papers (2020-03-10T21:39:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.