MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
- URL: http://arxiv.org/abs/2202.10583v1
- Date: Thu, 17 Feb 2022 13:37:35 GMT
- Title: MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned
- Authors: Anssi Kanervisto, Stephanie Milani, Karolis Ramanauskas, Nicholay
Topin, Zichuan Lin, Junyou Li, Jianing Shi, Deheng Ye, Qiang Fu, Wei Yang,
Weijun Hong, Zhongyue Huang, Haicheng Chen, Guangjun Zeng, Yue Lin, Vincent
Micheli, Eloi Alonso, Fran\c{c}ois Fleuret, Alexander Nikulin, Yury Belousov,
Oleg Svidchenko, Aleksei Shpilman
- Abstract summary: Reinforcement learning competitions advance the field by providing appropriate scope and support to develop solutions toward a specific problem.
We hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.
The participants of this easier track were able to obtain a diamond, and the participants of the harder track progressed the generalizable solutions in the same task.
- Score: 60.11039031794829
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning competitions advance the field by providing
appropriate scope and support to develop solutions toward a specific problem.
To promote the development of more broadly applicable methods, organizers need
to enforce the use of general techniques, the use of sample-efficient methods,
and the reproducibility of the results. While beneficial for the research
community, these restrictions come at a cost -- increased difficulty. If the
barrier for entry is too high, many potential participants are demoralized.
With this in mind, we hosted the third edition of the MineRL ObtainDiamond
competition, MineRL Diamond 2021, with a separate track in which we permitted
any solution to promote the participation of newcomers. With this track and
more extensive tutorials and support, we saw an increased number of
submissions. The participants of this easier track were able to obtain a
diamond, and the participants of the harder track progressed the generalizable
solutions in the same task.
Related papers
- Benchmarking Robustness and Generalization in Multi-Agent Systems: A
Case Study on Neural MMO [50.58083807719749]
We present the results of the second Neural MMO challenge, hosted at IJCAI 2022, which received 1600+ submissions.
This competition targets robustness and generalization in multi-agent systems.
We will open-source our benchmark including the environment wrapper, baselines, a visualization tool, and selected policies for further research.
arXiv Detail & Related papers (2023-08-30T07:16:11Z) - The STOIC2021 COVID-19 AI challenge: applying reusable training
methodologies to private data [60.94672667514737]
This study implements the Type Three (T3) challenge format, which allows for training solutions on private data.
With T3, challenge organizers train a provided by the participants on sequestered training data.
The winning solution obtained an area under the receiver operating characteristic curve for discerning between severe and non-severe COVID-19 of 0.815.
arXiv Detail & Related papers (2023-06-18T05:48:28Z) - Lessons learned from the NeurIPS 2021 MetaDL challenge: Backbone
fine-tuning without episodic meta-learning dominates for few-shot learning
image classification [40.901760230639496]
We describe the design of the MetaDL competition series, the datasets, the best experimental results, and the top-ranked methods in the NeurIPS 2021 challenge.
The solutions of the top participants have been open-sourced.
arXiv Detail & Related papers (2022-06-15T10:27:23Z) - Retrospective on the 2021 BASALT Competition on Learning from Human
Feedback [92.37243979045817]
The goal of the competition was to promote research towards agents that use learning from human feedback (LfHF) techniques to solve open-world tasks.
Rather than mandating the use of LfHF techniques, we described four tasks in natural language to be accomplished in the video game Minecraft.
Teams developed a diverse range of LfHF algorithms across a variety of possible human feedback types.
arXiv Detail & Related papers (2022-04-14T17:24:54Z) - Towards robust and domain agnostic reinforcement learning competitions [12.731614722371376]
Reinforcement learning competitions have formed the basis for standard research benchmarks.
Despite this, a majority of challenges suffer from the same fundamental problems.
We present a new framework of competition design that promotes the development of algorithms that overcome these barriers.
arXiv Detail & Related papers (2021-06-07T16:15:46Z) - The MineRL 2020 Competition on Sample Efficient Reinforcement Learning
using Human Priors [62.9301667732188]
We propose a second iteration of the MineRL Competition.
The primary goal of the competition is to foster the development of algorithms which can efficiently leverage human demonstrations.
The competition is structured into two rounds in which competitors are provided several paired versions of the dataset and environment.
At the end of each round, competitors submit containerized versions of their learning algorithms to the AIcrowd platform.
arXiv Detail & Related papers (2021-01-26T20:32:30Z) - Competing Bandits: The Perils of Exploration Under Competition [99.68537519404727]
We study the interplay between exploration and competition on online platforms.
We find that stark competition induces firms to commit to a "greedy" bandit algorithm that leads to low welfare.
We investigate two channels for weakening the competition: relaxing the rationality of users and giving one firm a first-mover advantage.
arXiv Detail & Related papers (2020-07-20T14:19:08Z) - Retrospective Analysis of the 2019 MineRL Competition on Sample
Efficient Reinforcement Learning [27.440055101691115]
We held the MineRL Competition on Sample Efficient Reinforcement Learning Using Human Priors at the Thirty-third Conference on Neural Information Processing Systems (NeurIPS)
The primary goal of this competition was to promote the development of algorithms that use human demonstrations alongside reinforcement learning to reduce the number of samples needed to solve complex, hierarchical, and sparse environments.
arXiv Detail & Related papers (2020-03-10T21:39:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.