Teaching Machines to Converse
- URL: http://arxiv.org/abs/2001.11701v1
- Date: Fri, 31 Jan 2020 08:28:07 GMT
- Title: Teaching Machines to Converse
- Authors: Jiwei Li
- Abstract summary: This dissertation attempts to tackle challenges presented by neural network models in open-domain dialogue generation systems.
We develop interactive question-answering dialogue systems by giving the agent the ability to ask questions and training a conversation agent through interactions with humans in an online fashion.
- Score: 24.64148203917298
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability of a machine to communicate with humans has long been associated
with the general success of AI. This dates back to Alan Turing's epoch-making
work in the early 1950s, which proposes that a machine's intelligence can be
tested by how well it, the machine, can fool a human into believing that the
machine is a human through dialogue conversations. Many systems learn
generation rules from a minimal set of authored rules or labels on top of
hand-coded rules or templates, and thus are both expensive and difficult to
extend to open-domain scenarios. Recently, the emergence of neural network
models the potential to solve many of the problems in dialogue learning that
earlier systems cannot tackle: the end-to-end neural frameworks offer the
promise of scalability and language-independence, together with the ability to
track the dialogue state and then mapping between states and dialogue actions
in a way not possible with conventional systems. On the other hand, neural
systems bring about new challenges: they tend to output dull and generic
responses; they lack a consistent or a coherent persona; they are usually
optimized through single-turn conversations and are incapable of handling the
long-term success of a conversation; and they are not able to take the
advantage of the interactions with humans. This dissertation attempts to tackle
these challenges: Contributions are two-fold: (1) we address new challenges
presented by neural network models in open-domain dialogue generation systems;
(2) we develop interactive question-answering dialogue systems by (a) giving
the agent the ability to ask questions and (b) training a conversation agent
through interactions with humans in an online fashion, where a bot improves
through communicating with humans and learning from the mistakes that it makes.
Related papers
- Interactive Dialogue Agents via Reinforcement Learning on Hindsight Regenerations [58.65755268815283]
Many real dialogues are interactive, meaning an agent's utterances will influence their conversational partner, elicit information, or change their opinion.
We use this fact to rewrite and augment existing suboptimal data, and train via offline reinforcement learning (RL) an agent that outperforms both prompting and learning from unaltered human demonstrations.
Our results in a user study with real humans show that our approach greatly outperforms existing state-of-the-art dialogue agents.
arXiv Detail & Related papers (2024-11-07T21:37:51Z) - Enabling High-Level Machine Reasoning with Cognitive Neuro-Symbolic
Systems [67.01132165581667]
We propose to enable high-level reasoning in AI systems by integrating cognitive architectures with external neuro-symbolic components.
We illustrate a hybrid framework centered on ACT-R and we discuss the role of generative models in recent and future applications.
arXiv Detail & Related papers (2023-11-13T21:20:17Z) - PK-Chat: Pointer Network Guided Knowledge Driven Generative Dialogue
Model [79.64376762489164]
PK-Chat is a Pointer network guided generative dialogue model, incorporating a unified pretrained language model and a pointer network over knowledge graphs.
The words generated by PK-Chat in the dialogue are derived from the prediction of word lists and the direct prediction of the external knowledge graph knowledge.
Based on the PK-Chat, a dialogue system is built for academic scenarios in the case of geosciences.
arXiv Detail & Related papers (2023-04-02T18:23:13Z) - From Knowledge Augmentation to Multi-tasking: Towards Human-like
Dialogue Systems [0.0]
The goal of building dialogue agents that can converse with humans naturally has been a long-standing dream of researchers.
In this thesis, we focus on methods that address the numerous issues that have been imposing the gap between artificial conversational agents and human-level interlocutors.
arXiv Detail & Related papers (2022-11-14T17:27:07Z) - Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning [35.67318830455459]
We develop a real-time, open-ended dialogue system that uses reinforcement learning (RL) to power a bot's conversational skill at scale.
Our work pairs the succinct embedding of the conversation state generated using SOTA (supervised) language models with RL techniques that are particularly suited to a dynamic action space.
arXiv Detail & Related papers (2022-07-25T16:12:33Z) - A Review of Dialogue Systems: From Trained Monkeys to Stochastic Parrots [0.0]
We aim to deploy artificial intelligence to build automated dialogue agents that can converse with humans.
We present a broad overview of methods developed to build dialogue systems over the years.
arXiv Detail & Related papers (2021-11-02T08:07:55Z) - Retrieval Augmentation Reduces Hallucination in Conversation [49.35235945543833]
We explore the use of neural-retrieval-in-the-loop architectures for knowledge-grounded dialogue.
We show that our best models obtain state-of-the-art performance on two knowledge-grounded conversational tasks.
arXiv Detail & Related papers (2021-04-15T16:24:43Z) - Learning Adaptive Language Interfaces through Decomposition [89.21937539950966]
We introduce a neural semantic parsing system that learns new high-level abstractions through decomposition.
Users interactively teach the system by breaking down high-level utterances describing novel behavior into low-level steps.
arXiv Detail & Related papers (2020-10-11T08:27:07Z) - Knowledge Injection into Dialogue Generation via Language Models [85.65843021510521]
InjK is a two-stage approach to inject knowledge into a dialogue generation model.
First, we train a large-scale language model and query it as textual knowledge.
Second, we frame a dialogue generation model to sequentially generate textual knowledge and a corresponding response.
arXiv Detail & Related papers (2020-04-30T07:31:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.