Beyond the Granularity: Multi-Perspective Dialogue Collaborative
Selection for Dialogue State Tracking
- URL: http://arxiv.org/abs/2205.10059v1
- Date: Fri, 20 May 2022 10:08:45 GMT
- Title: Beyond the Granularity: Multi-Perspective Dialogue Collaborative
Selection for Dialogue State Tracking
- Authors: Jinyu Guo, Kai Shuang, Jijie Li, Zihan Wang and Yixuan Liu
- Abstract summary: In dialogue state tracking, dialogue history is a crucial material, and its utilization varies between different models.
We propose DiCoS-DST to dynamically select the relevant dialogue contents corresponding to each slot for state updating.
Our approach achieves new state-of-the-art performance on MultiWOZ 2.1 and MultiWOZ 2.2, and achieves superior performance on multiple mainstream benchmark datasets.
- Score: 18.172993687706708
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In dialogue state tracking, dialogue history is a crucial material, and its
utilization varies between different models. However, no matter how the
dialogue history is used, each existing model uses its own consistent dialogue
history during the entire state tracking process, regardless of which slot is
updated. Apparently, it requires different dialogue history to update different
slots in different turns. Therefore, using consistent dialogue contents may
lead to insufficient or redundant information for different slots, which
affects the overall performance. To address this problem, we devise DiCoS-DST
to dynamically select the relevant dialogue contents corresponding to each slot
for state updating. Specifically, it first retrieves turn-level utterances of
dialogue history and evaluates their relevance to the slot from a combination
of three perspectives: (1) its explicit connection to the slot name; (2) its
relevance to the current turn dialogue; (3) Implicit Mention Oriented
Reasoning. Then these perspectives are combined to yield a decision, and only
the selected dialogue contents are fed into State Generator, which explicitly
minimizes the distracting information passed to the downstream state
prediction. Experimental results show that our approach achieves new
state-of-the-art performance on MultiWOZ 2.1 and MultiWOZ 2.2, and achieves
superior performance on multiple mainstream benchmark datasets (including
Sim-M, Sim-R, and DSTC2).
Related papers
- Enhancing Visual Dialog State Tracking through Iterative Object-Entity Alignment in Multi-Round Conversations [3.784841749866846]
We introduce Multi-round Dialogue State Tracking model (MDST)
MDST captures each round of dialog history, constructing internal dialogue state representations defined as 2-tuples of vision-language representations.
Experimental results on the VisDial v1.0 dataset demonstrate that MDST achieves a new state-of-the-art performance in generative setting.
arXiv Detail & Related papers (2024-08-13T08:36:15Z) - Multi-turn Dialogue Comprehension from a Topic-aware Perspective [70.37126956655985]
This paper proposes to model multi-turn dialogues from a topic-aware perspective.
We use a dialogue segmentation algorithm to split a dialogue passage into topic-concentrated fragments in an unsupervised way.
We also present a novel model, Topic-Aware Dual-Attention Matching (TADAM) Network, which takes topic segments as processing elements.
arXiv Detail & Related papers (2023-09-18T11:03:55Z) - Act-Aware Slot-Value Predicting in Multi-Domain Dialogue State Tracking [5.816391291790977]
Dialogue state tracking (DST) aims to track human-machine interactions and generate state representations for managing the dialogue.
Recent advances in machine reading comprehension predict both categorical and non-categorical types of slots for dialogue state tracking.
We formulate and incorporate dialogue acts, and leverage recent advances in machine reading comprehension to predict both categorical and non-categorical types of slots for dialogue state tracking.
arXiv Detail & Related papers (2022-08-04T05:18:30Z) - Multimodal Dialogue State Tracking [97.25466640240619]
Video-Dialogue Transformer Network (VDTN) learns contextual dependencies between videos and dialogues to generate multimodal dialogue states.
VDTN combines both object-level features and segment-level features and learns contextual dependencies between videos and dialogues to generate multimodal dialogue states.
arXiv Detail & Related papers (2022-06-16T03:18:42Z) - Back to the Future: Bidirectional Information Decoupling Network for
Multi-turn Dialogue Modeling [80.51094098799736]
We propose Bidirectional Information Decoupling Network (BiDeN) as a universal dialogue encoder.
BiDeN explicitly incorporates both the past and future contexts and can be generalized to a wide range of dialogue-related tasks.
Experimental results on datasets of different downstream tasks demonstrate the universality and effectiveness of our BiDeN.
arXiv Detail & Related papers (2022-04-18T03:51:46Z) - Dialogue State Tracking with Multi-Level Fusion of Predicted Dialogue
States and Conversations [2.6529642559155944]
We propose the Dialogue State Tracking with Multi-Level Fusion of Predicted Dialogue States and Conversations network.
This model extracts information of each dialogue turn by modeling interactions among each turn utterance, the corresponding last dialogue states, and dialogue slots.
arXiv Detail & Related papers (2021-07-12T02:30:30Z) - Domain State Tracking for a Simplified Dialogue System [3.962145079528281]
We present DoTS, a task-oriented dialogue system that uses a simplified input context instead of the entire dialogue history.
DoTS improves the inform rate and success rate by 1.09 points and 1.24 points, respectively, compared to the previous state-of-the-art model on MultiWOZ.
arXiv Detail & Related papers (2021-03-11T13:00:54Z) - Rethinking Dialogue State Tracking with Reasoning [76.0991910623001]
This paper proposes to track dialogue states gradually with reasoning over dialogue turns with the help of the back-end data.
Empirical results demonstrate that our method significantly outperforms the state-of-the-art methods by 38.6% in terms of joint belief accuracy for MultiWOZ 2.1.
arXiv Detail & Related papers (2020-05-27T02:05:33Z) - UniConv: A Unified Conversational Neural Architecture for Multi-domain
Task-oriented Dialogues [101.96097419995556]
"UniConv" is a novel unified neural architecture for end-to-end conversational systems in task-oriented dialogues.
We conduct comprehensive experiments in dialogue state tracking, context-to-text, and end-to-end settings on the MultiWOZ2.1 benchmark.
arXiv Detail & Related papers (2020-04-29T16:28:22Z) - Non-Autoregressive Dialog State Tracking [122.2328875457225]
We propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST)
NADST can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots.
Our results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus.
arXiv Detail & Related papers (2020-02-19T06:39:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.