DialoGen: Generalized Long-Range Context Representation for Dialogue
Systems
- URL: http://arxiv.org/abs/2210.06282v4
- Date: Tue, 3 Oct 2023 05:17:54 GMT
- Title: DialoGen: Generalized Long-Range Context Representation for Dialogue
Systems
- Authors: Suvodip Dey, Maunendra Sankar Desarkar, Asif Ekbal, P.K. Srijith
- Abstract summary: We propose DialoGen, a novel framework for dialogue generation with a generalized context representation.
We study the effectiveness of our proposed method on both dialogue generation (open-domain) and understanding (DST)
- Score: 36.23733762476647
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Long-range context modeling is crucial to both dialogue understanding and
generation. The most popular method for dialogue context representation is to
concatenate the last-$k$ utterances in chronological order. However, this
method may not be ideal for conversations containing long-range dependencies,
i.e., when there is a need to look beyond last-$k$ utterances to generate a
meaningful response. In this work, we propose DialoGen, a novel encoder-decoder
based framework for dialogue generation with a generalized context
representation that can look beyond the last-$k$ utterances. The main idea of
our approach is to identify and utilize the most relevant historical utterances
instead of last-$k$, which also enables the compact representation of dialogue
history with fewer tokens. We study the effectiveness of our proposed method on
both dialogue generation (open-domain) and understanding (DST). Even with a
compact context representation, DialoGen performs comparably to the
state-of-the-art models on the open-domain DailyDialog dataset. We observe a
similar behavior on the DST task of the MultiWOZ dataset when the proposed
context representation is applied to existing DST models. We also discuss the
generalizability and interpretability of DialoGen and show that the relevance
score of previous utterances agrees well with human cognition.
Related papers
- Enhancing Visual Dialog State Tracking through Iterative Object-Entity Alignment in Multi-Round Conversations [3.784841749866846]
We introduce Multi-round Dialogue State Tracking model (MDST)
MDST captures each round of dialog history, constructing internal dialogue state representations defined as 2-tuples of vision-language representations.
Experimental results on the VisDial v1.0 dataset demonstrate that MDST achieves a new state-of-the-art performance in generative setting.
arXiv Detail & Related papers (2024-08-13T08:36:15Z) - Multi-grained Hypergraph Interest Modeling for Conversational
Recommendation [75.65483522949857]
We propose a novel multi-grained hypergraph interest modeling approach to capture user interest beneath intricate historical data.
In our approach, we first employ the hypergraph structure to model users' historical dialogue sessions and form a session-based hypergraph, which captures coarse-grained, session-level relations.
We further conduct multi-grained hypergraph convolution on the two kinds of hypergraphs, and utilize the enhanced representations to develop interest-aware CRS.
arXiv Detail & Related papers (2023-05-04T13:13:44Z) - Unsupervised Dialogue Topic Segmentation with Topic-aware Utterance
Representation [51.22712675266523]
Dialogue Topic (DTS) plays an essential role in a variety of dialogue modeling tasks.
We propose a novel unsupervised DTS framework, which learns topic-aware utterance representations from unlabeled dialogue data.
arXiv Detail & Related papers (2023-05-04T11:35:23Z) - X-ReCoSa: Multi-Scale Context Aggregation For Multi-Turn Dialogue
Generation [0.0]
In multi-turn dialogue generation, responses are not only related to the topic and background of the context but also related to words and phrases in the sentences of the context.
Currently widely used hierarchical dialog models solely rely on context representations from the utterance-level encoder, ignoring the sentence representations output by the word-level encoder.
We propose a new dialog model X-ReCoSa to tackle this problem which aggregates multi-scale context information for hierarchical dialog models.
arXiv Detail & Related papers (2023-03-14T12:15:52Z) - History-Aware Hierarchical Transformer for Multi-session Open-domain
Dialogue System [59.78425104243993]
We propose History-Aware Hierarchical Transformer (HAHT) for multi-session open-domain dialogue.
HAHT maintains a long-term memory of history conversations and utilizes history information to understand current conversation context.
Experimental results on a large-scale Multi-Session Conversation dataset suggest that the proposed HAHT model consistently outperforms baseline models.
arXiv Detail & Related papers (2023-02-02T06:54:33Z) - Dialogue Meaning Representation for Task-Oriented Dialogue Systems [51.91615150842267]
We propose Dialogue Meaning Representation (DMR), a flexible and easily extendable representation for task-oriented dialogue.
Our representation contains a set of nodes and edges with inheritance hierarchy to represent rich semantics for compositional semantics and task-specific concepts.
We propose two evaluation tasks to evaluate different machine learning based dialogue models, and further propose a novel coreference resolution model GNNCoref for the graph-based coreference resolution task.
arXiv Detail & Related papers (2022-04-23T04:17:55Z) - Precognition in Task-oriented Dialogue Understanding: Posterior
Regularization by Future Context [8.59600111891194]
We propose to jointly model historical and future information through the posterior regularization method.
We optimize the KL distance between these to regularize our model during training.
Experiments on two dialogue datasets validate the effectiveness of our proposed method.
arXiv Detail & Related papers (2022-03-07T09:58:50Z) - CREDIT: Coarse-to-Fine Sequence Generation for Dialogue State Tracking [44.38388988238695]
A dialogue state tracker aims to accurately find a compact representation of the current dialogue status.
We employ a structured state representation and cast dialogue state tracking as a sequence generation problem.
Experiments demonstrate our tracker achieves encouraging joint goal accuracy for the five domains in MultiWOZ 2.0 and MultiWOZ 2.1 datasets.
arXiv Detail & Related papers (2020-09-22T10:27:18Z) - Ranking Enhanced Dialogue Generation [77.8321855074999]
How to effectively utilize the dialogue history is a crucial problem in multi-turn dialogue generation.
Previous works usually employ various neural network architectures to model the history.
This paper proposes a Ranking Enhanced Dialogue generation framework.
arXiv Detail & Related papers (2020-08-13T01:49:56Z) - Modeling Long Context for Task-Oriented Dialogue State Generation [51.044300192906995]
We propose a multi-task learning model with a simple yet effective utterance tagging technique and a bidirectional language model.
Our approaches attempt to solve the problem that the performance of the baseline significantly drops when the input dialogue context sequence is long.
In our experiments, our proposed model achieves a 7.03% relative improvement over the baseline, establishing a new state-of-the-art joint goal accuracy of 52.04% on the MultiWOZ 2.0 dataset.
arXiv Detail & Related papers (2020-04-29T11:02:25Z) - Efficient Context and Schema Fusion Networks for Multi-Domain Dialogue
State Tracking [32.36259992245]
Dialogue state tracking (DST) aims at estimating the current dialogue state given all the preceding conversation.
For multi-domain DST, the data sparsity problem is a major obstacle due to increased numbers of state candidates and dialogue lengths.
We utilize the previous dialogue state (predicted) and the current dialogue utterance as the input for DST.
arXiv Detail & Related papers (2020-04-07T13:46:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.