Modeling Worlds in Text
- URL: http://arxiv.org/abs/2106.09578v1
- Date: Thu, 17 Jun 2021 15:02:16 GMT
- Title: Modeling Worlds in Text
- Authors: Prithviraj Ammanabrolu, Mark O. Riedl
- Abstract summary: We provide a dataset that enables the creation of learning agents that can build knowledge graph-based world models of interactive narratives.
Our dataset provides 24198 mappings between rich natural language observations and knowledge graphs.
The training data is collected across 27 games in multiple genres and contains a further 7836 heldout instances over 9 additional games in the test set.
- Score: 16.67845396797253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We provide a dataset that enables the creation of learning agents that can
build knowledge graph-based world models of interactive narratives. Interactive
narratives -- or text-adventure games -- are partially observable environments
structured as long puzzles or quests in which an agent perceives and interacts
with the world purely through textual natural language. Each individual game
typically contains hundreds of locations, characters, and objects -- each with
their own unique descriptions -- providing an opportunity to study the problem
of giving language-based agents the structured memory necessary to operate in
such worlds. Our dataset provides 24198 mappings between rich natural language
observations and: (1) knowledge graphs that reflect the world state in the form
of a map; (2) natural language actions that are guaranteed to cause a change in
that particular world state. The training data is collected across 27 games in
multiple genres and contains a further 7836 heldout instances over 9 additional
games in the test set. We further provide baseline models using rules-based,
question-answering, and sequence learning approaches in addition to an analysis
of the data and corresponding learning tasks.
Related papers
- Visually Grounded Language Learning: a review of language games,
datasets, tasks, and models [60.2604624857992]
Many Vision+Language (V+L) tasks have been defined with the aim of creating models that can ground symbols in the visual modality.
In this work, we provide a systematic literature review of several tasks and models proposed in the V+L field.
arXiv Detail & Related papers (2023-12-05T02:17:29Z) - Learning to Model the World with Language [100.76069091703505]
To interact with humans and act in the world, agents need to understand the range of language that people use and relate it to the visual world.
Our key idea is that agents should interpret such diverse language as a signal that helps them predict the future.
We instantiate this in Dynalang, an agent that learns a multimodal world model to predict future text and image representations.
arXiv Detail & Related papers (2023-07-31T17:57:49Z) - Foundational Models Defining a New Era in Vision: A Survey and Outlook [151.49434496615427]
Vision systems to see and reason about the compositional nature of visual scenes are fundamental to understanding our world.
The models learned to bridge the gap between such modalities coupled with large-scale training data facilitate contextual reasoning, generalization, and prompt capabilities at test time.
The output of such models can be modified through human-provided prompts without retraining, e.g., segmenting a particular object by providing a bounding box, having interactive dialogues by asking questions about an image or video scene or manipulating the robot's behavior through language instructions.
arXiv Detail & Related papers (2023-07-25T17:59:18Z) - Tachikuma: Understading Complex Interactions with Multi-Character and
Novel Objects by Large Language Models [67.20964015591262]
We introduce a benchmark named Tachikuma, comprising a Multiple character and novel Object based interaction Estimation task and a supporting dataset.
The dataset captures log data from real-time communications during gameplay, providing diverse, grounded, and complex interactions for further explorations.
We present a simple prompting baseline and evaluate its performance, demonstrating its effectiveness in enhancing interaction understanding.
arXiv Detail & Related papers (2023-07-24T07:40:59Z) - Infusing Commonsense World Models with Graph Knowledge [89.27044249858332]
We study the setting of generating narratives in an open world text adventure game.
A graph representation of the underlying game state can be used to train models that consume and output both grounded graph representations and natural language descriptions and actions.
arXiv Detail & Related papers (2023-01-13T19:58:27Z) - Learning Knowledge Graph-based World Models of Textual Environments [16.67845396797253]
This work focuses on the task of building world models of text-based game environments.
Our world model learns to simultaneously: (1) predict changes in the world caused by an agent's actions when representing the world as a knowledge graph; and (2) generate the set of contextually relevant natural language actions required to operate in the world.
arXiv Detail & Related papers (2021-06-17T15:45:54Z) - Situated Language Learning via Interactive Narratives [16.67845396797253]
This paper explores the question of how to imbue learning agents with the ability to understand and generate contextually relevant natural language.
Two key components in creating such agents are interactivity and environment grounding.
We discuss the unique challenges a text games' puzzle-like structure combined with natural language state-and-action spaces provides.
arXiv Detail & Related papers (2021-03-18T01:55:16Z) - Learning Dynamic Belief Graphs to Generalize on Text-Based Games [55.59741414135887]
Playing text-based games requires skills in processing natural language and sequential decision making.
In this work, we investigate how an agent can plan and generalize in text-based games using graph-structured representations learned end-to-end from raw text.
arXiv Detail & Related papers (2020-02-21T04:38:37Z) - Bringing Stories Alive: Generating Interactive Fiction Worlds [19.125250090589397]
We focus on procedurally generating interactive fiction worlds that players "see" and "talk to" using natural language.
We present a method that first extracts a partial knowledge graph encoding basic information regarding world structure.
This knowledge graph is then automatically completed utilizing thematic knowledge and used to guide a neural language generation model.
arXiv Detail & Related papers (2020-01-28T04:13:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.