Context-Aware Interaction Network for Question Matching
- URL: http://arxiv.org/abs/2104.08451v1
- Date: Sat, 17 Apr 2021 05:03:56 GMT
- Title: Context-Aware Interaction Network for Question Matching
- Authors: Zhe Hu, Zuohui Fu, Yu Yin, Gerard de Melo and Cheng Peng
- Abstract summary: We propose a context-aware interaction network (COIN) to align two sequences and infer their semantic relationship.
Specifically, each interaction block includes (1) a context-aware cross-attention mechanism to effectively integrate contextual information, and (2) a gate fusion layer to flexibly interpolate aligned representations.
- Score: 51.76812857301819
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Impressive milestones have been achieved in text matching by adopting a
cross-attention mechanism to capture pertinent semantic connections between two
sentences. However, these cross-attention mechanisms focus on word-level links
between the two inputs, neglecting the importance of contextual information. We
propose a context-aware interaction network (COIN) to properly align two
sequences and infer their semantic relationship. Specifically, each interaction
block includes (1) a context-aware cross-attention mechanism to effectively
integrate contextual information, and (2) a gate fusion layer to flexibly
interpolate aligned representations. We apply multiple stacked interaction
blocks to produce alignments at different levels and gradually refine the
attention results. Experiments on two question matching datasets and detailed
analyses confirm the effectiveness of our model.
Related papers
- Multi-Grained Multimodal Interaction Network for Entity Linking [65.30260033700338]
Multimodal entity linking task aims at resolving ambiguous mentions to a multimodal knowledge graph.
We propose a novel Multi-GraIned Multimodal InteraCtion Network $textbf(MIMIC)$ framework for solving the MEL task.
arXiv Detail & Related papers (2023-07-19T02:11:19Z) - A Hierarchical Interactive Network for Joint Span-based Aspect-Sentiment
Analysis [34.1489054082536]
We propose a hierarchical interactive network (HI-ASA) to model two-way interactions between two tasks appropriately.
We use cross-stitch mechanism to combine the different task-specific features selectively as the input to ensure proper two-way interactions.
Experiments on three real-world datasets demonstrate HI-ASA's superiority over baselines.
arXiv Detail & Related papers (2022-08-24T03:03:49Z) - Semantic Role Aware Correlation Transformer for Text to Video Retrieval [23.183653281610866]
This paper proposes a novel transformer that explicitly disentangles the text and video into semantic roles of objects, spatial contexts and temporal contexts.
Preliminary results on popular YouCook2 indicate that our approach surpasses a current state-of-the-art method, with a high margin in all metrics.
arXiv Detail & Related papers (2022-06-26T11:28:03Z) - Towards Joint Intent Detection and Slot Filling via Higher-order
Attention [47.78365472691051]
Intent detection (ID) and Slot filling (SF) are two major tasks in spoken language understanding (SLU)
We propose a Bilinear attention block, which exploits both the contextual and channel-wise bilinear attention distributions.
We show that our approach yields improvements compared with the state-of-the-art approach.
arXiv Detail & Related papers (2021-09-18T09:50:23Z) - Co-GAT: A Co-Interactive Graph Attention Network for Joint Dialog Act
Recognition and Sentiment Classification [34.711179589196355]
In a dialog system, dialog act recognition and sentiment classification are two correlative tasks.
We propose a Co-Interactive Graph Attention Network (Co-GAT) to jointly perform the two tasks.
Experimental results on two public datasets show that our model successfully captures the two sources of information.
arXiv Detail & Related papers (2020-12-24T14:10:24Z) - Interactive Fusion of Multi-level Features for Compositional Activity
Recognition [100.75045558068874]
We present a novel framework that accomplishes this goal by interactive fusion.
We implement the framework in three steps, namely, positional-to-appearance feature extraction, semantic feature interaction, and semantic-to-positional prediction.
We evaluate our approach on two action recognition datasets, Something-Something and Charades.
arXiv Detail & Related papers (2020-12-10T14:17:18Z) - DCR-Net: A Deep Co-Interactive Relation Network for Joint Dialog Act
Recognition and Sentiment Classification [77.59549450705384]
In dialog system, dialog act recognition and sentiment classification are two correlative tasks.
Most of the existing systems either treat them as separate tasks or just jointly model the two tasks.
We propose a Deep Co-Interactive Relation Network (DCR-Net) to explicitly consider the cross-impact and model the interaction between the two tasks.
arXiv Detail & Related papers (2020-08-16T14:13:32Z) - Adaptive Interaction Fusion Networks for Fake News Detection [12.459707991466168]
We propose Adaptive Interaction Fusion Networks (AIFN) to fulfill cross-interaction fusion among features for fake news detection.
To establish feature associations, we devise semantic-level fusion self-attention networks (SFSN)
Experiments on two real-world datasets, RumourEval and PHEME, demonstrate that AIFN achieves the state-of-the-art performance and boosts accuracy by more than 2.05% and 1.90%, respectively.
arXiv Detail & Related papers (2020-04-21T13:51:03Z) - Cascaded Human-Object Interaction Recognition [175.60439054047043]
We introduce a cascade architecture for a multi-stage, coarse-to-fine HOI understanding.
At each stage, an instance localization network progressively refines HOI proposals and feeds them into an interaction recognition network.
With our carefully-designed human-centric relation features, these two modules work collaboratively towards effective interaction understanding.
arXiv Detail & Related papers (2020-03-09T17:05:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.