I love your chain mail! Making knights smile in a fantasy game world:
Open-domain goal-oriented dialogue agents
- URL: http://arxiv.org/abs/2002.02878v2
- Date: Mon, 10 Feb 2020 20:45:20 GMT
- Title: I love your chain mail! Making knights smile in a fantasy game world:
Open-domain goal-oriented dialogue agents
- Authors: Shrimai Prabhumoye and Margaret Li and Jack Urbanek and Emily Dinan
and Douwe Kiela and Jason Weston and Arthur Szlam
- Abstract summary: We train a goal-oriented model with reinforcement learning against an imitation-learned chit-chat'' model.
We show that both models outperform an inverse model baseline and can converse naturally with their dialogue partner in order to achieve goals.
- Score: 69.68400056148336
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dialogue research tends to distinguish between chit-chat and goal-oriented
tasks. While the former is arguably more naturalistic and has a wider use of
language, the latter has clearer metrics and a straightforward learning signal.
Humans effortlessly combine the two, for example engaging in chit-chat with the
goal of exchanging information or eliciting a specific response. Here, we
bridge the divide between these two domains in the setting of a rich
multi-player text-based fantasy environment where agents and humans engage in
both actions and dialogue. Specifically, we train a goal-oriented model with
reinforcement learning against an imitation-learned ``chit-chat'' model with
two approaches: the policy either learns to pick a topic or learns to pick an
utterance given the top-K utterances from the chit-chat model. We show that
both models outperform an inverse model baseline and can converse naturally
with their dialogue partner in order to achieve goals.
Related papers
- Searching for Snippets of Open-Domain Dialogue in Task-Oriented Dialogue
Datasets [0.0]
chit-chat/opendomain dialogues focus on holding a socially engaging talk with a user.
Task-oriented dialogues portray functional goals, such as making a restaurant reservation or booking a plane ticket.
Our study shows that sequences related to social talk are indeed naturally present, motivating further research on ways chitchat is combined into task-oriented dialogues.
arXiv Detail & Related papers (2023-11-23T16:08:39Z) - Zero-Shot Goal-Directed Dialogue via RL on Imagined Conversations [70.7884839812069]
Large language models (LLMs) have emerged as powerful and general solutions to many natural language tasks.
However, many of the most important applications of language generation are interactive, where an agent has to talk to a person to reach a desired outcome.
In this work, we explore a new method for adapting LLMs with RL for such goal-directed dialogue.
arXiv Detail & Related papers (2023-11-09T18:45:16Z) - Multi-turn Dialogue Comprehension from a Topic-aware Perspective [70.37126956655985]
This paper proposes to model multi-turn dialogues from a topic-aware perspective.
We use a dialogue segmentation algorithm to split a dialogue passage into topic-concentrated fragments in an unsupervised way.
We also present a novel model, Topic-Aware Dual-Attention Matching (TADAM) Network, which takes topic segments as processing elements.
arXiv Detail & Related papers (2023-09-18T11:03:55Z) - Multi-Party Chat: Conversational Agents in Group Settings with Humans
and Models [39.80729604768669]
We evaluate the ability of language models to act as one or more characters in multi-party conversations.
We find that our new dataset, MultiLIGHT, can help bring significant improvements in the group setting.
arXiv Detail & Related papers (2023-04-26T21:41:17Z) - SalesBot: Transitioning from Chit-Chat to Task-Oriented Dialogues [22.89699254073016]
How smoothly transitioning from social chatting to task-oriented dialogues is important for triggering business opportunities.
This paper proposes a framework to automatically generate many dialogues without human involvement.
The released data has a great potential of guiding future research directions and commercial activities.
arXiv Detail & Related papers (2022-04-22T09:31:13Z) - Few-shot Language Coordination by Modeling Theory of Mind [95.54446989205117]
We study the task of few-shot $textitlanguage coordination$.
We require the lead agent to coordinate with a $textitpopulation$ of agents with different linguistic abilities.
This requires the ability to model the partner's beliefs, a vital component of human communication.
arXiv Detail & Related papers (2021-07-12T19:26:11Z) - Adding Chit-Chat to Enhance Task-Oriented Dialogues [36.93917437554091]
Chit-Chat can be added to task-oriented dialogues to make virtual assistant conversations more engaging and interactive.
We present our new chit-chat-based annotations to 23.8K dialogues from two popular task-oriented dialogue datasets.
We also propose three new models for adding chit-chat to task-oriented dialogues, explicitly trained to predict user goals and to generate contextually relevant chit-chat responses.
arXiv Detail & Related papers (2020-10-24T03:22:43Z) - How to Motivate Your Dragon: Teaching Goal-Driven Agents to Speak and
Act in Fantasy Worlds [47.7511759322784]
We seek to create agents that both act and communicate with other agents in pursuit of a goal.
We introduce a reinforcement learning system that incorporates large-scale language modeling-based and commonsense reasoning-based pre-training.
We conduct zero-shot evaluations using held-out human expert demonstrations, showing that our agents are able to act consistently and talk naturally with respect to their motivations.
arXiv Detail & Related papers (2020-10-01T21:06:21Z) - Knowledge Injection into Dialogue Generation via Language Models [85.65843021510521]
InjK is a two-stage approach to inject knowledge into a dialogue generation model.
First, we train a large-scale language model and query it as textual knowledge.
Second, we frame a dialogue generation model to sequentially generate textual knowledge and a corresponding response.
arXiv Detail & Related papers (2020-04-30T07:31:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.