End to End Dialogue Transformer
- URL: http://arxiv.org/abs/2008.10392v1
- Date: Mon, 24 Aug 2020 12:43:08 GMT
- Title: End to End Dialogue Transformer
- Authors: Ond\v{r}ej M\v{e}kota, Memduh G\"ok{\i}rmak, Petr Laitoch
- Abstract summary: We are inspired by the performance of the recurrent neural network-based model Sequicity.
We propose a dialogue system based on the Transformer architecture instead of Sequicity's RNN-based architecture.
- Score: 0.0019832631155284838
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dialogue systems attempt to facilitate conversations between humans and
computers, for purposes as diverse as small talk to booking a vacation. We are
here inspired by the performance of the recurrent neural network-based model
Sequicity, which when conducting a dialogue uses a sequence-to-sequence
architecture to first produce a textual representation of what is going on in
the dialogue, and in a further step use this along with database findings to
produce a reply to the user. We here propose a dialogue system based on the
Transformer architecture instead of Sequicity's RNN-based architecture, that
works similarly in an end-to-end, sequence-to-sequence fashion.
Related papers
- A Static and Dynamic Attention Framework for Multi Turn Dialogue Generation [37.79563028123686]
In open domain multi turn dialogue generation, it is essential to modeling the contextual semantics of the dialogue history.
Previous research had verified the effectiveness of the hierarchical recurrent encoder-decoder framework on open domain multi turn dialogue generation.
We propose a static and dynamic attention-based approach to model the dialogue history and then generate open domain multi turn dialogue responses.
arXiv Detail & Related papers (2024-10-28T06:05:34Z) - Dialog-to-Actions: Building Task-Oriented Dialogue System via
Action-Level Generation [7.110201160927713]
We propose a task-oriented dialogue system via action-level generation.
Specifically, we first construct dialogue actions from large-scale dialogues and represent each natural language (NL) response as a sequence of dialogue actions.
We train a Sequence-to-Sequence model which takes the dialogue history as input and outputs sequence of dialogue actions.
arXiv Detail & Related papers (2023-04-03T11:09:20Z) - Manual-Guided Dialogue for Flexible Conversational Agents [84.46598430403886]
How to build and use dialogue data efficiently, and how to deploy models in different domains at scale can be critical issues in building a task-oriented dialogue system.
We propose a novel manual-guided dialogue scheme, where the agent learns the tasks from both dialogue and manuals.
Our proposed scheme reduces the dependence of dialogue models on fine-grained domain ontology, and makes them more flexible to adapt to various domains.
arXiv Detail & Related papers (2022-08-16T08:21:12Z) - Back to the Future: Bidirectional Information Decoupling Network for
Multi-turn Dialogue Modeling [80.51094098799736]
We propose Bidirectional Information Decoupling Network (BiDeN) as a universal dialogue encoder.
BiDeN explicitly incorporates both the past and future contexts and can be generalized to a wide range of dialogue-related tasks.
Experimental results on datasets of different downstream tasks demonstrate the universality and effectiveness of our BiDeN.
arXiv Detail & Related papers (2022-04-18T03:51:46Z) - Structure Extraction in Task-Oriented Dialogues with Slot Clustering [94.27806592467537]
In task-oriented dialogues, dialogue structure has often been considered as transition graphs among dialogue states.
We propose a simple yet effective approach for structure extraction in task-oriented dialogues.
arXiv Detail & Related papers (2022-02-28T20:18:12Z) - Discovering Dialog Structure Graph for Open-Domain Dialog Generation [51.29286279366361]
We conduct unsupervised discovery of dialog structure from chitchat corpora.
We then leverage it to facilitate dialog generation in downstream systems.
We present a Discrete Variational Auto-Encoder with Graph Neural Network (DVAE-GNN), to discover a unified human-readable dialog structure.
arXiv Detail & Related papers (2020-12-31T10:58:37Z) - Hierarchical Transformer for Task Oriented Dialog Systems [11.743662338418867]
We show how a standard transformer can be morphed into any hierarchical encoder, including HRED and HIBERT like models, by using specially designed attention masks and positional encodings.
We demonstrate that Hierarchical Hierarchical helps achieve better natural language understanding of the contexts in transformer-based models for task-oriented dialog systems.
arXiv Detail & Related papers (2020-10-24T10:08:52Z) - Structured Attention for Unsupervised Dialogue Structure Induction [110.12561786644122]
We propose to incorporate structured attention layers into a Variational Recurrent Neural Network (VRNN) model with discrete latent states to learn dialogue structure in an unsupervised fashion.
Compared to a vanilla VRNN, structured attention enables a model to focus on different parts of the source sentence embeddings while enforcing a structural inductive bias.
arXiv Detail & Related papers (2020-09-17T23:07:03Z) - Ranking Enhanced Dialogue Generation [77.8321855074999]
How to effectively utilize the dialogue history is a crucial problem in multi-turn dialogue generation.
Previous works usually employ various neural network architectures to model the history.
This paper proposes a Ranking Enhanced Dialogue generation framework.
arXiv Detail & Related papers (2020-08-13T01:49:56Z) - Exploring Recurrent, Memory and Attention Based Architectures for
Scoring Interactional Aspects of Human-Machine Text Dialog [9.209192502526285]
This paper builds on previous work in this direction to investigate multiple neural architectures.
We conduct experiments on a conversational database of text dialogs from human learners interacting with a cloud-based dialog system.
We find that fusion of multiple architectures performs competently on our automated scoring task relative to expert inter-rater agreements.
arXiv Detail & Related papers (2020-05-20T03:23:00Z) - EmpTransfo: A Multi-head Transformer Architecture for Creating
Empathetic Dialog Systems [4.41738804598711]
This paper presents EmpTransfo, a multi-head Transformer architecture for creating an empathetic dialog system.
We show that utilizing the history of emotions and other metadata can improve the quality of generated conversations.
arXiv Detail & Related papers (2020-03-05T23:09:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.