DCR-Net: A Deep Co-Interactive Relation Network for Joint Dialog Act
Recognition and Sentiment Classification
- URL: http://arxiv.org/abs/2008.06914v1
- Date: Sun, 16 Aug 2020 14:13:32 GMT
- Title: DCR-Net: A Deep Co-Interactive Relation Network for Joint Dialog Act
Recognition and Sentiment Classification
- Authors: Libo Qin, Wanxiang Che, Yangming Li, Minheng Ni, Ting Liu
- Abstract summary: In dialog system, dialog act recognition and sentiment classification are two correlative tasks.
Most of the existing systems either treat them as separate tasks or just jointly model the two tasks.
We propose a Deep Co-Interactive Relation Network (DCR-Net) to explicitly consider the cross-impact and model the interaction between the two tasks.
- Score: 77.59549450705384
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In dialog system, dialog act recognition and sentiment classification are two
correlative tasks to capture speakers intentions, where dialog act and
sentiment can indicate the explicit and the implicit intentions separately.
Most of the existing systems either treat them as separate tasks or just
jointly model the two tasks by sharing parameters in an implicit way without
explicitly modeling mutual interaction and relation. To address this problem,
we propose a Deep Co-Interactive Relation Network (DCR-Net) to explicitly
consider the cross-impact and model the interaction between the two tasks by
introducing a co-interactive relation layer. In addition, the proposed relation
layer can be stacked to gradually capture mutual knowledge with multiple steps
of interaction. Especially, we thoroughly study different relation layers and
their effects. Experimental results on two public datasets (Mastodon and
Dailydialog) show that our model outperforms the state-of-the-art joint model
by 4.3% and 3.4% in terms of F1 score on dialog act recognition task, 5.7% and
12.4% on sentiment classification respectively. Comprehensive analysis
empirically verifies the effectiveness of explicitly modeling the relation
between the two tasks and the multi-steps interaction mechanism. Finally, we
employ the Bidirectional Encoder Representation from Transformer (BERT) in our
framework, which can further boost our performance in both tasks.
Related papers
- A Bi-directional Multi-hop Inference Model for Joint Dialog Sentiment
Classification and Act Recognition [25.426172735931463]
The joint task of Dialog Sentiment Classification (DSC) and Act Recognition (DAR) aims to predict the sentiment label and act label for each utterance in a dialog simultaneously.
We propose a Bi-directional Multi-hop Inference Model (BMIM) that iteratively extract and integrate rich sentiment and act clues in a bi-directional manner.
BMIM outperforms state-of-the-art baselines by at least 2.6% on F1 score in DAR and 1.4% on F1 score in DSC.
arXiv Detail & Related papers (2023-08-08T17:53:24Z) - Multi-Grained Multimodal Interaction Network for Entity Linking [65.30260033700338]
Multimodal entity linking task aims at resolving ambiguous mentions to a multimodal knowledge graph.
We propose a novel Multi-GraIned Multimodal InteraCtion Network $textbf(MIMIC)$ framework for solving the MEL task.
arXiv Detail & Related papers (2023-07-19T02:11:19Z) - A Hierarchical Interactive Network for Joint Span-based Aspect-Sentiment
Analysis [34.1489054082536]
We propose a hierarchical interactive network (HI-ASA) to model two-way interactions between two tasks appropriately.
We use cross-stitch mechanism to combine the different task-specific features selectively as the input to ensure proper two-way interactions.
Experiments on three real-world datasets demonstrate HI-ASA's superiority over baselines.
arXiv Detail & Related papers (2022-08-24T03:03:49Z) - Disentangled Representation Learning for Text-Video Retrieval [51.861423831566626]
Cross-modality interaction is a critical component in Text-Video Retrieval (TVR)
We study the interaction paradigm in depth, where we find that its computation can be split into two terms.
We propose a disentangled framework to capture a sequential and hierarchical representation.
arXiv Detail & Related papers (2022-03-14T13:55:33Z) - Hierarchical Modeling for Out-of-Scope Domain and Intent Classification [55.23920796595698]
This paper focuses on out-of-scope intent classification in dialog systems.
We propose a hierarchical multi-task learning approach based on a joint model to classify domain and intent simultaneously.
Experiments show that the model outperforms existing methods in terms of accuracy, out-of-scope recall and F1.
arXiv Detail & Related papers (2021-04-30T06:38:23Z) - Co-GAT: A Co-Interactive Graph Attention Network for Joint Dialog Act
Recognition and Sentiment Classification [34.711179589196355]
In a dialog system, dialog act recognition and sentiment classification are two correlative tasks.
We propose a Co-Interactive Graph Attention Network (Co-GAT) to jointly perform the two tasks.
Experimental results on two public datasets show that our model successfully captures the two sources of information.
arXiv Detail & Related papers (2020-12-24T14:10:24Z) - A Co-Interactive Transformer for Joint Slot Filling and Intent Detection [61.109486326954205]
Intent detection and slot filling are two main tasks for building a spoken language understanding (SLU) system.
Previous studies either model the two tasks separately or only consider the single information flow from intent to slot.
We propose a Co-Interactive Transformer to consider the cross-impact between the two tasks simultaneously.
arXiv Detail & Related papers (2020-10-08T10:16:52Z) - Cascaded Human-Object Interaction Recognition [175.60439054047043]
We introduce a cascade architecture for a multi-stage, coarse-to-fine HOI understanding.
At each stage, an instance localization network progressively refines HOI proposals and feeds them into an interaction recognition network.
With our carefully-designed human-centric relation features, these two modules work collaboratively towards effective interaction understanding.
arXiv Detail & Related papers (2020-03-09T17:05:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.