WordCraft: An Environment for Benchmarking Commonsense Agents
- URL: http://arxiv.org/abs/2007.09185v1
- Date: Fri, 17 Jul 2020 18:40:46 GMT
- Title: WordCraft: An Environment for Benchmarking Commonsense Agents
- Authors: Minqi Jiang, Jelena Luketina, Nantas Nardelli, Pasquale Minervini,
Philip H. S. Torr, Shimon Whiteson, Tim Rockt\"aschel
- Abstract summary: We propose WordCraft, an RL environment based on Little Alchemy 2.
This lightweight environment is fast to run and built upon entities and relations inspired by real-world semantics.
- Score: 107.20421897619002
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability to quickly solve a wide range of real-world tasks requires a
commonsense understanding of the world. Yet, how to best extract such knowledge
from natural language corpora and integrate it with reinforcement learning (RL)
agents remains an open challenge. This is partly due to the lack of lightweight
simulation environments that sufficiently reflect the semantics of the real
world and provide knowledge sources grounded with respect to observations in an
RL environment. To better enable research on agents making use of commonsense
knowledge, we propose WordCraft, an RL environment based on Little Alchemy 2.
This lightweight environment is fast to run and built upon entities and
relations inspired by real-world semantics. We evaluate several representation
learning methods on this new benchmark and propose a new method for integrating
knowledge graphs with an RL agent.
Related papers
- Improving Sample Efficiency of Reinforcement Learning with Background Knowledge from Large Language Models [33.504700578933424]
Low sample efficiency is an enduring challenge of reinforcement learning (RL)
We introduce a framework that harnesses large language models to extract background knowledge of an environment.
Our experiments show that these methods achieve significant sample efficiency improvements in a spectrum of downstream tasks.
arXiv Detail & Related papers (2024-07-04T14:33:47Z) - Language Guided Exploration for RL Agents in Text Environments [15.256908785183617]
Large Language Models (LLMs), with a wealth of world knowledge, can help RL agents learn quickly and adapt to distribution shifts.
We introduce Language Guided Exploration (LGE) framework, which uses a pre-trained language model to provide decision-level guidance to an RL agent (called EXPLORER)
We observe that on ScienceWorld, a challenging text environment, LGE outperforms vanilla RL agents significantly and also outperforms other sophisticated methods like Behaviour Cloning and Text Decision Transformer.
arXiv Detail & Related papers (2024-03-05T17:26:41Z) - Vision-Language Models Provide Promptable Representations for Reinforcement Learning [67.40524195671479]
We propose a novel approach that uses the vast amounts of general and indexable world knowledge encoded in vision-language models (VLMs) pre-trained on Internet-scale data for embodied reinforcement learning (RL)
We show that our approach can use chain-of-thought prompting to produce representations of common-sense semantic reasoning, improving policy performance in novel scenes by 1.5 times.
arXiv Detail & Related papers (2024-02-05T00:48:56Z) - Towards an Information Theoretic Framework of Context-Based Offline
Meta-Reinforcement Learning [50.976910714839065]
Context-based OMRL (COMRL) as a popular paradigm, aims to learn a universal policy conditioned on effective task representations.
We show that COMRL algorithms are essentially optimizing the same mutual information objective between the task variable $boldsymbolM$ and its latent representation $boldsymbolZ$ by implementing various approximate bounds.
Based on the theoretical insight and the information bottleneck principle, we arrive at a novel algorithm dubbed UNICORN, which exhibits remarkable generalization across a broad spectrum of RL benchmarks.
arXiv Detail & Related papers (2024-02-04T09:58:42Z) - Beyond Factuality: A Comprehensive Evaluation of Large Language Models
as Knowledge Generators [78.63553017938911]
Large language models (LLMs) outperform information retrieval techniques for downstream knowledge-intensive tasks.
However, community concerns abound regarding the factuality and potential implications of using this uncensored knowledge.
We introduce CONNER, designed to evaluate generated knowledge from six important perspectives.
arXiv Detail & Related papers (2023-10-11T08:22:37Z) - Brain in a Vat: On Missing Pieces Towards Artificial General
Intelligence in Large Language Models [83.63242931107638]
We propose four characteristics of generally intelligent agents.
We argue that active engagement with objects in the real world delivers more robust signals for forming conceptual representations.
We conclude by outlining promising future research directions in the field of artificial general intelligence.
arXiv Detail & Related papers (2023-07-07T13:58:16Z) - Ghost in the Minecraft: Generally Capable Agents for Open-World
Environments via Large Language Models with Text-based Knowledge and Memory [97.87093169454431]
Ghost in the Minecraft (GITM) is a novel framework that integrates Large Language Models (LLMs) with text-based knowledge and memory.
We develop a set of structured actions and leverage LLMs to generate action plans for the agents to execute.
The resulting LLM-based agent markedly surpasses previous methods, achieving a remarkable improvement of +47.5% in success rate.
arXiv Detail & Related papers (2023-05-25T17:59:49Z) - Leveraging class abstraction for commonsense reinforcement learning via
residual policy gradient methods [23.199881381599617]
We propose to use the subclass relationships present in open-source knowledge graphs to abstract away from specific objects.
We develop a residual policy gradient method that is able to integrate knowledge across different abstraction levels in the class hierarchy.
arXiv Detail & Related papers (2022-01-28T14:03:49Z) - REIN-2: Giving Birth to Prepared Reinforcement Learning Agents Using
Reinforcement Learning Agents [0.0]
In this paper, we introduce a meta-learning scheme that shifts the objective of learning to solve a task into the objective of learning to learn to solve a task (or a set of tasks)
Our model, named REIN-2, is a meta-learning scheme formulated within the RL framework, the goal of which is to develop a meta-RL agent that learns how to produce other RL agents.
Compared to traditional state-of-the-art Deep RL algorithms, experimental results show remarkable performance of our model in popular OpenAI Gym environments.
arXiv Detail & Related papers (2021-10-11T10:13:49Z) - Text-based RL Agents with Commonsense Knowledge: New Challenges,
Environments and Baselines [40.03754436370682]
We show that agents which incorporate commonsense knowledge in TextWorld Commonsense perform better, while acting more efficiently.
We conduct user-studies to estimate human performance on TWC and show that there is ample room for future improvement.
arXiv Detail & Related papers (2020-10-08T06:20:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.