Towards Solving Fuzzy Tasks with Human Feedback: A Retrospective of the
MineRL BASALT 2022 Competition
- URL: http://arxiv.org/abs/2303.13512v1
- Date: Thu, 23 Mar 2023 17:59:17 GMT
- Title: Towards Solving Fuzzy Tasks with Human Feedback: A Retrospective of the
MineRL BASALT 2022 Competition
- Authors: Stephanie Milani, Anssi Kanervisto, Karolis Ramanauskas, Sander
Schulhoff, Brandon Houghton, Sharada Mohanty, Byron Galbraith, Ke Chen, Yan
Song, Tianze Zhou, Bingquan Yu, He Liu, Kai Guan, Yujing Hu, Tangjie Lv,
Federico Malato, Florian Leopold, Amogh Raut, Ville Hautam\"aki, Andrew
Melnik, Shu Ishida, Jo\~ao F. Henriques, Robert Klassert, Walter Laurito,
Ellen Novoseller, Vinicius G. Goecks, Nicholas Waytowich, David Watkins, Josh
Miller, Rohin Shah
- Abstract summary: The BASALT challenge asks teams to compete to develop algorithms to solve tasks with hard-to-specify reward functions in Minecraft.
We describe the competition and provide an overview of the top solutions.
- Score: 20.922425732605756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To facilitate research in the direction of fine-tuning foundation models from
human feedback, we held the MineRL BASALT Competition on Fine-Tuning from Human
Feedback at NeurIPS 2022. The BASALT challenge asks teams to compete to develop
algorithms to solve tasks with hard-to-specify reward functions in Minecraft.
Through this competition, we aimed to promote the development of algorithms
that use human feedback as channels to learn the desired behavior. We describe
the competition and provide an overview of the top solutions. We conclude by
discussing the impact of the competition and future directions for improvement.
Related papers
- AI Olympics challenge with Evolutionary Soft Actor Critic [5.076263094490715]
Our solution is based on a Model-free Deep Reinforcement Learning approach combined with an evolutionary strategy.
We will briefly describe the algorithms that have been used and then provide details of the approach.
arXiv Detail & Related papers (2024-09-02T09:34:18Z) - The NeurIPS 2022 Neural MMO Challenge: A Massively Multiagent
Competition with Specialization and Trade [41.639843908635875]
The NeurIPS-2022 Neural MMO Challenge attracted 500 participants and received over 1,600 submissions.
This year's competition runs on the latest v1.6 Neural MMO, which introduces new equipment, combat, trading, and a better scoring system.
This paper summarizes the design and results of the challenge, explores the potential of this environment as a benchmark for learning methods, and presents some practical reinforcement learning approaches for complex tasks with sparse rewards.
arXiv Detail & Related papers (2023-11-07T04:14:45Z) - Benchmarking Robustness and Generalization in Multi-Agent Systems: A
Case Study on Neural MMO [50.58083807719749]
We present the results of the second Neural MMO challenge, hosted at IJCAI 2022, which received 1600+ submissions.
This competition targets robustness and generalization in multi-agent systems.
We will open-source our benchmark including the environment wrapper, baselines, a visualization tool, and selected policies for further research.
arXiv Detail & Related papers (2023-08-30T07:16:11Z) - RangL: A Reinforcement Learning Competition Platform [82.1944886411643]
RangL aims to encourage the wider uptake of reinforcement learning by supporting competitions relating to real-world dynamic decision problems.
This article describes the reusable code repository developed by the RangL team and deployed for the 2022 Pathways to Net Zero Challenge.
The winning solutions to this particular Challenge seek to optimize the UK's energy transition policy to net zero carbon emissions by 2050.
arXiv Detail & Related papers (2022-07-28T09:44:21Z) - Retrospective on the 2021 BASALT Competition on Learning from Human
Feedback [92.37243979045817]
The goal of the competition was to promote research towards agents that use learning from human feedback (LfHF) techniques to solve open-world tasks.
Rather than mandating the use of LfHF techniques, we described four tasks in natural language to be accomplished in the video game Minecraft.
Teams developed a diverse range of LfHF algorithms across a variety of possible human feedback types.
arXiv Detail & Related papers (2022-04-14T17:24:54Z) - MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned [60.11039031794829]
Reinforcement learning competitions advance the field by providing appropriate scope and support to develop solutions toward a specific problem.
We hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.
The participants of this easier track were able to obtain a diamond, and the participants of the harder track progressed the generalizable solutions in the same task.
arXiv Detail & Related papers (2022-02-17T13:37:35Z) - The MineRL 2020 Competition on Sample Efficient Reinforcement Learning
using Human Priors [62.9301667732188]
We propose a second iteration of the MineRL Competition.
The primary goal of the competition is to foster the development of algorithms which can efficiently leverage human demonstrations.
The competition is structured into two rounds in which competitors are provided several paired versions of the dataset and environment.
At the end of each round, competitors submit containerized versions of their learning algorithms to the AIcrowd platform.
arXiv Detail & Related papers (2021-01-26T20:32:30Z) - Retrospective Analysis of the 2019 MineRL Competition on Sample
Efficient Reinforcement Learning [27.440055101691115]
We held the MineRL Competition on Sample Efficient Reinforcement Learning Using Human Priors at the Thirty-third Conference on Neural Information Processing Systems (NeurIPS)
The primary goal of this competition was to promote the development of algorithms that use human demonstrations alongside reinforcement learning to reduce the number of samples needed to solve complex, hierarchical, and sparse environments.
arXiv Detail & Related papers (2020-03-10T21:39:52Z) - Analysing Affective Behavior in the First ABAW 2020 Competition [49.90617840789334]
The Affective Behavior Analysis in-the-wild (ABAW) 2020 Competition is the first Competition aiming at automatic analysis of the three main behavior tasks.
We describe this Competition, to be held in conjunction with the IEEE Conference on Face and Gesture Recognition, May 2020, in Buenos Aires, Argentina.
We outline the evaluation metrics, present both the baseline system and the top-3 performing teams' methodologies per Challenge and finally present their obtained results.
arXiv Detail & Related papers (2020-01-30T15:41:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.