I Cast Detect Thoughts: Learning to Converse and Guide with Intents and
Theory-of-Mind in Dungeons and Dragons
- URL: http://arxiv.org/abs/2212.10060v2
- Date: Tue, 30 May 2023 23:09:37 GMT
- Title: I Cast Detect Thoughts: Learning to Converse and Guide with Intents and
Theory-of-Mind in Dungeons and Dragons
- Authors: Pei Zhou, Andrew Zhu, Jennifer Hu, Jay Pujara, Xiang Ren, Chris
Callison-Burch, Yejin Choi, Prithviraj Ammanabrolu
- Abstract summary: We study teacher-student natural language interactions in a goal-driven environment in Dungeons and Dragons.
Our approach is to decompose and model these interactions into (1) the Dungeon Master's intent to guide players toward a given goal; (2) the DM's guidance utterance to the players expressing this intent; and (3) a theory-of-mind (ToM) model that anticipates the players' reaction to the guidance one turn into the future.
- Score: 82.28503603235364
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a novel task, G4C, to study teacher-student natural language
interactions in a goal-driven and grounded environment. Dungeons and Dragons
(D&D), a role-playing game, provides an ideal setting to investigate such
interactions. Here, the Dungeon Master (DM), i.e., the teacher, guides the
actions of several players -- students, each with their own personas and
abilities -- to achieve shared goals grounded in a fantasy world. Our approach
is to decompose and model these interactions into (1) the DM's intent to guide
players toward a given goal; (2) the DM's guidance utterance to the players
expressing this intent; and (3) a theory-of-mind (ToM) model that anticipates
the players' reaction to the guidance one turn into the future. We develop a
novel reinforcement learning (RL) method for training a DM that generates
guidance for players by rewarding utterances where the intent matches the
ToM-anticipated player actions. Human and automated evaluations show that a DM
trained to explicitly model intents and incorporate ToM of the players using RL
generates better-quality guidance that is 3x more likely to fulfill the DM's
intent than a vanilla natural language generation (NLG) approach.
Related papers
- Attacks and Defenses for Generative Diffusion Models: A Comprehensive Survey [5.300811350105823]
Diffusion models (DMs) have achieved state-of-the-art performance on various generative tasks.
Recent studies have shown that DMs are prone to a wide range of attacks.
arXiv Detail & Related papers (2024-08-06T18:52:17Z) - Skill Check: Some Considerations on the Evaluation of Gamemastering
Models for Role-playing Games [0.5803309695504829]
In role-playing games a Game Master (GM) is the player in charge of the game, who must design the challenges the players face and narrate the outcomes of their actions.
In this work we discuss some challenges to model GMs from an Interactive Storytelling and Natural Language Processing perspective.
Following those challenges we propose three test categories to evaluate such dialogue systems, and we use them to test ChatGPT, Bard and OpenAssistant as out-of-the-box GMs.
arXiv Detail & Related papers (2023-09-24T17:19:36Z) - CALYPSO: LLMs as Dungeon Masters' Assistants [46.61924662589895]
Large language models (LLMs) have shown remarkable abilities to generate coherent natural language text.
We introduce CALYPSO, a system of LLM-powered interfaces that support DMs with information and inspiration specific to their own scenario.
When given access to CALYPSO, DMs reported that it generated high-fidelity text suitable for direct presentation to players, and low-fidelity ideas that the DM could develop further while maintaining their creative agency.
arXiv Detail & Related papers (2023-08-15T02:57:00Z) - Dungeons and Dragons as a Dialog Challenge for Artificial Intelligence [28.558934742150022]
We frame D&D as a dialogue system challenge, where the tasks are to both generate the next conversational turn in the game and predict the state of the game given the dialogue history.
We create a gameplay dataset consisting of nearly 900 games, with a total of 7,000 players, 800,000 dialogue turns, 500,000 dice rolls, and 58 million words.
We train a large language model (LM) to generate the next game turn, conditioning it on different information.
arXiv Detail & Related papers (2022-10-13T15:43:39Z) - A Mixture-of-Expert Approach to RL-based Dialogue Management [56.08449336469477]
We use reinforcement learning to develop a dialogue agent that avoids being short-sighted (outputting generic utterances) and maximizes overall user satisfaction.
Most existing RL approaches to DM train the agent at the word-level, and thus, have to deal with aly complex action space even for a medium-size vocabulary.
We develop a RL-based DM using a novel mixture of expert language model (MoE-LM) that consists of (i) a LM capable of learning diverse semantics for conversation histories, (ii) a number of specialized LMs (or experts) capable of generating utterances corresponding to a
arXiv Detail & Related papers (2022-05-31T19:00:41Z) - DeepCrawl: Deep Reinforcement Learning for Turn-based Strategy Games [137.86426963572214]
We introduce DeepCrawl, a fully-playable Roguelike prototype for iOS and Android in which all agents are controlled by policy networks trained using Deep Reinforcement Learning (DRL)
Our aim is to understand whether recent advances in DRL can be used to develop convincing behavioral models for non-player characters in videogames.
arXiv Detail & Related papers (2020-12-03T13:53:29Z) - How to Motivate Your Dragon: Teaching Goal-Driven Agents to Speak and
Act in Fantasy Worlds [47.7511759322784]
We seek to create agents that both act and communicate with other agents in pursuit of a goal.
We introduce a reinforcement learning system that incorporates large-scale language modeling-based and commonsense reasoning-based pre-training.
We conduct zero-shot evaluations using held-out human expert demonstrations, showing that our agents are able to act consistently and talk naturally with respect to their motivations.
arXiv Detail & Related papers (2020-10-01T21:06:21Z) - I love your chain mail! Making knights smile in a fantasy game world:
Open-domain goal-oriented dialogue agents [69.68400056148336]
We train a goal-oriented model with reinforcement learning against an imitation-learned chit-chat'' model.
We show that both models outperform an inverse model baseline and can converse naturally with their dialogue partner in order to achieve goals.
arXiv Detail & Related papers (2020-02-07T16:22:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.