Retrospective Analysis of the 2019 MineRL Competition on Sample
Efficient Reinforcement Learning
- URL: http://arxiv.org/abs/2003.05012v4
- Date: Thu, 18 Jun 2020 16:54:23 GMT
- Title: Retrospective Analysis of the 2019 MineRL Competition on Sample
Efficient Reinforcement Learning
- Authors: Stephanie Milani, Nicholay Topin, Brandon Houghton, William H. Guss,
Sharada P. Mohanty, Keisuke Nakata, Oriol Vinyals, Noboru Sean Kuno
- Abstract summary: We held the MineRL Competition on Sample Efficient Reinforcement Learning Using Human Priors at the Thirty-third Conference on Neural Information Processing Systems (NeurIPS)
The primary goal of this competition was to promote the development of algorithms that use human demonstrations alongside reinforcement learning to reduce the number of samples needed to solve complex, hierarchical, and sparse environments.
- Score: 27.440055101691115
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To facilitate research in the direction of sample efficient reinforcement
learning, we held the MineRL Competition on Sample Efficient Reinforcement
Learning Using Human Priors at the Thirty-third Conference on Neural
Information Processing Systems (NeurIPS 2019). The primary goal of this
competition was to promote the development of algorithms that use human
demonstrations alongside reinforcement learning to reduce the number of samples
needed to solve complex, hierarchical, and sparse environments. We describe the
competition, outlining the primary challenge, the competition design, and the
resources that we provided to the participants. We provide an overview of the
top solutions, each of which use deep reinforcement learning and/or imitation
learning. We also discuss the impact of our organizational decisions on the
competition and future directions for improvement.
Related papers
- Take the Bull by the Horns: Hard Sample-Reweighted Continual Training
Improves LLM Generalization [165.98557106089777]
A key challenge is to enhance the capabilities of large language models (LLMs) amid a looming shortage of high-quality training data.
Our study starts from an empirical strategy for the light continual training of LLMs using their original pre-training data sets.
We then formalize this strategy into a principled framework of Instance-Reweighted Distributionally Robust Optimization.
arXiv Detail & Related papers (2024-02-22T04:10:57Z) - CompeteSMoE -- Effective Training of Sparse Mixture of Experts via
Competition [52.2034494666179]
Sparse mixture of experts (SMoE) offers an appealing solution to scale up the model complexity beyond the mean of increasing the network's depth or width.
We propose a competition mechanism to address this fundamental challenge of representation collapse.
By routing inputs only to experts with the highest neural response, we show that, under mild assumptions, competition enjoys the same convergence rate as the optimal estimator.
arXiv Detail & Related papers (2024-02-04T15:17:09Z) - Benchmarking Robustness and Generalization in Multi-Agent Systems: A
Case Study on Neural MMO [50.58083807719749]
We present the results of the second Neural MMO challenge, hosted at IJCAI 2022, which received 1600+ submissions.
This competition targets robustness and generalization in multi-agent systems.
We will open-source our benchmark including the environment wrapper, baselines, a visualization tool, and selected policies for further research.
arXiv Detail & Related papers (2023-08-30T07:16:11Z) - Lessons learned from the NeurIPS 2021 MetaDL challenge: Backbone
fine-tuning without episodic meta-learning dominates for few-shot learning
image classification [40.901760230639496]
We describe the design of the MetaDL competition series, the datasets, the best experimental results, and the top-ranked methods in the NeurIPS 2021 challenge.
The solutions of the top participants have been open-sourced.
arXiv Detail & Related papers (2022-06-15T10:27:23Z) - Retrospective on the 2021 BASALT Competition on Learning from Human
Feedback [92.37243979045817]
The goal of the competition was to promote research towards agents that use learning from human feedback (LfHF) techniques to solve open-world tasks.
Rather than mandating the use of LfHF techniques, we described four tasks in natural language to be accomplished in the video game Minecraft.
Teams developed a diverse range of LfHF algorithms across a variety of possible human feedback types.
arXiv Detail & Related papers (2022-04-14T17:24:54Z) - MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned [60.11039031794829]
Reinforcement learning competitions advance the field by providing appropriate scope and support to develop solutions toward a specific problem.
We hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.
The participants of this easier track were able to obtain a diamond, and the participants of the harder track progressed the generalizable solutions in the same task.
arXiv Detail & Related papers (2022-02-17T13:37:35Z) - Towards robust and domain agnostic reinforcement learning competitions [12.731614722371376]
Reinforcement learning competitions have formed the basis for standard research benchmarks.
Despite this, a majority of challenges suffer from the same fundamental problems.
We present a new framework of competition design that promotes the development of algorithms that overcome these barriers.
arXiv Detail & Related papers (2021-06-07T16:15:46Z) - The MineRL 2020 Competition on Sample Efficient Reinforcement Learning
using Human Priors [62.9301667732188]
We propose a second iteration of the MineRL Competition.
The primary goal of the competition is to foster the development of algorithms which can efficiently leverage human demonstrations.
The competition is structured into two rounds in which competitors are provided several paired versions of the dataset and environment.
At the end of each round, competitors submit containerized versions of their learning algorithms to the AIcrowd platform.
arXiv Detail & Related papers (2021-01-26T20:32:30Z) - Sample Efficient Reinforcement Learning through Learning from
Demonstrations in Minecraft [4.3952888284140785]
We show how human demonstrations can improve final performance of agents on the Minecraft minigame ObtainDiamond with only 8M frames of environment interaction.
Our solution placed 3rd in the NeurIPS MineRL Competition for Sample-Efficient Reinforcement Learning.
arXiv Detail & Related papers (2020-03-12T23:46:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.