Learning to Memorize Entailment and Discourse Relations for
Persona-Consistent Dialogues
- URL: http://arxiv.org/abs/2301.04871v1
- Date: Thu, 12 Jan 2023 08:37:00 GMT
- Title: Learning to Memorize Entailment and Discourse Relations for
Persona-Consistent Dialogues
- Authors: Ruijun Chen, Jin Wang, Liang-Chih Yu and Xuejie Zhang
- Abstract summary: Existing works have improved the performance of dialogue systems by intentionally learning interlocutor personas with sophisticated network structures.
This study proposes a method of learning to memorize entailment and discourse relations for persona-consistent dialogue tasks.
- Score: 8.652711997920463
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Maintaining engagement and consistency is particularly important in dialogue
systems. Existing works have improved the performance of dialogue systems by
intentionally learning interlocutor personas with sophisticated network
structures. One issue with this approach is that it requires more personal
corpora with annotations. Additionally, these models typically perform the next
utterance prediction to generate a response but neglect the discourse coherence
in the entire conversation. To address these issues, this study proposes a
method of learning to memorize entailment and discourse relations for
persona-consistent dialogue tasks. Entailment text pairs in natural language
inference dataset were applied to learn latent entailment relations as external
memories by premise-to-hypothesis generation task. Furthermore, an internal
memory with a similar architecture was applied to the discourse information in
the dialogue. Placing orthogonality restrictions on these two memory spaces
ensures that the latent entailment relations remain dialogue-independent. Both
memories collaborate to obtain entailment and discourse representation for the
generation, allowing a deeper understanding of both consistency and coherence.
Experiments on two large public datasets, PersonaChat and DSTC7-AVSD,
demonstrated the effectiveness of the proposed method. Both automatic and human
evaluations indicate that the proposed model outperforms several strong
baselines in terms of both persona consistency and response coherence. Our
source code is available at https://github.com/Chenrj233/LMEDR.
Related papers
- Enhancing Personality Recognition in Dialogue by Data Augmentation and
Heterogeneous Conversational Graph Networks [30.33718960981521]
Personality recognition is useful for enhancing robots' ability to tailor user-adaptive responses.
One of the challenges in this task is a limited number of speakers in existing dialogue corpora.
arXiv Detail & Related papers (2024-01-11T12:27:33Z) - Revisiting Conversation Discourse for Dialogue Disentanglement [88.3386821205896]
We propose enhancing dialogue disentanglement by taking full advantage of the dialogue discourse characteristics.
We develop a structure-aware framework to integrate the rich structural features for better modeling the conversational semantic context.
Our work has great potential to facilitate broader multi-party multi-thread dialogue applications.
arXiv Detail & Related papers (2023-06-06T19:17:47Z) - Pre-training Multi-party Dialogue Models with Latent Discourse Inference [85.9683181507206]
We pre-train a model that understands the discourse structure of multi-party dialogues, namely, to whom each utterance is replying.
To fully utilize the unlabeled data, we propose to treat the discourse structures as latent variables, then jointly infer them and pre-train the discourse-aware model.
arXiv Detail & Related papers (2023-05-24T14:06:27Z) - Dial2vec: Self-Guided Contrastive Learning of Unsupervised Dialogue
Embeddings [41.79937481022846]
We introduce the task of learning unsupervised dialogue embeddings.
Trivial approaches such as combining pre-trained word or sentence embeddings and encoding through pre-trained language models have been shown to be feasible.
We propose a self-guided contrastive learning approach named dial2vec.
arXiv Detail & Related papers (2022-10-27T11:14:06Z) - TOD-DA: Towards Boosting the Robustness of Task-oriented Dialogue
Modeling on Spoken Conversations [24.245354500835465]
We propose a novel model-agnostic data augmentation paradigm to boost the robustness of task-oriented dialogue modeling on spoken conversations.
Our approach ranked first in both tasks of DSTC10 Track2, a benchmark for task-oriented dialogue modeling on spoken conversations.
arXiv Detail & Related papers (2021-12-23T10:04:25Z) - Who says like a style of Vitamin: Towards Syntax-Aware
DialogueSummarization using Multi-task Learning [2.251583286448503]
We focus on the association between utterances from individual speakers and unique syntactic structures.
Speakers have unique textual styles that can contain linguistic information, such as voiceprint.
We employ multi-task learning of both syntax-aware information and dialogue summarization.
arXiv Detail & Related papers (2021-09-29T05:30:39Z) - "How Robust r u?": Evaluating Task-Oriented Dialogue Systems on Spoken
Conversations [87.95711406978157]
This work presents a new benchmark on spoken task-oriented conversations.
We study multi-domain dialogue state tracking and knowledge-grounded dialogue modeling.
Our data set enables speech-based benchmarking of task-oriented dialogue systems.
arXiv Detail & Related papers (2021-09-28T04:51:04Z) - Structural Pre-training for Dialogue Comprehension [51.215629336320305]
We present SPIDER, Structural Pre-traIned DialoguE Reader, to capture dialogue exclusive features.
To simulate the dialogue-like features, we propose two training objectives in addition to the original LM objectives.
Experimental results on widely used dialogue benchmarks verify the effectiveness of the newly introduced self-supervised tasks.
arXiv Detail & Related papers (2021-05-23T15:16:54Z) - Dialogue History Matters! Personalized Response Selectionin Multi-turn
Retrieval-based Chatbots [62.295373408415365]
We propose a personalized hybrid matching network (PHMN) for context-response matching.
Our contributions are two-fold: 1) our model extracts personalized wording behaviors from user-specific dialogue history as extra matching information.
We evaluate our model on two large datasets with user identification, i.e., personalized dialogue Corpus Ubuntu (P- Ubuntu) and personalized Weibo dataset (P-Weibo)
arXiv Detail & Related papers (2021-03-17T09:42:11Z) - Dialogue-Based Relation Extraction [53.2896545819799]
We present the first human-annotated dialogue-based relation extraction (RE) dataset DialogRE.
We argue that speaker-related information plays a critical role in the proposed task, based on an analysis of similarities and differences between dialogue-based and traditional RE tasks.
Experimental results demonstrate that a speaker-aware extension on the best-performing model leads to gains in both the standard and conversational evaluation settings.
arXiv Detail & Related papers (2020-04-17T03:51:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.