Short Text Classification via Knowledge powered Attention with
Similarity Matrix based CNN
- URL: http://arxiv.org/abs/2002.03350v2
- Date: Sun, 28 Feb 2021 23:40:10 GMT
- Title: Short Text Classification via Knowledge powered Attention with
Similarity Matrix based CNN
- Authors: Mingchen Li and Gabtone.Clinton and Yijia Miao and Feng Gao
- Abstract summary: We propose a knowledge powered attention with similarity matrix based convolutional neural network (KASM) model.
We use knowledge graph (KG) to enrich the semantic representation of short text, specially, the information of parent-entity is introduced in our model.
For the purpose of measuring the importance of knowledge, we introduce the attention mechanisms to choose the important information.
- Score: 6.6723692875904375
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Short text is becoming more and more popular on the web, such as Chat
Message, SMS and Product Reviews. Accurately classifying short text is an
important and challenging task. A number of studies have difficulties in
addressing this problem because of the word ambiguity and data sparsity. To
address this issue, we propose a knowledge powered attention with similarity
matrix based convolutional neural network (KASM) model, which can compute
comprehensive information by utilizing the knowledge and deep neural network.
We use knowledge graph (KG) to enrich the semantic representation of short
text, specially, the information of parent-entity is introduced in our model.
Meanwhile, we consider the word interaction in the literal-level between short
text and the representation of label, and utilize similarity matrix based
convolutional neural network (CNN) to extract it. For the purpose of measuring
the importance of knowledge, we introduce the attention mechanisms to choose
the important information. Experimental results on five standard datasets show
that our model significantly outperforms state-of-the-art methods.
Related papers
- KETM:A Knowledge-Enhanced Text Matching method [0.0]
We introduce a new model for text matching called the Knowledge Enhanced Text Matching model (KETM)
We use Wiktionary to retrieve the text word definitions as our external knowledge.
We fuse text and knowledge using a gating mechanism to learn the ratio of text and knowledge fusion.
arXiv Detail & Related papers (2023-08-11T17:08:14Z) - Neuro-Symbolic Artificial Intelligence (AI) for Intent based Semantic
Communication [85.06664206117088]
6G networks must consider semantics and effectiveness (at end-user) of the data transmission.
NeSy AI is proposed as a pillar for learning causal structure behind the observed data.
GFlowNet is leveraged for the first time in a wireless system to learn the probabilistic structure which generates the data.
arXiv Detail & Related papers (2022-05-22T07:11:57Z) - Hierarchical Heterogeneous Graph Representation Learning for Short Text
Classification [60.233529926965836]
We propose a new method called SHINE, which is based on graph neural network (GNN) for short text classification.
First, we model the short text dataset as a hierarchical heterogeneous graph consisting of word-level component graphs.
Then, we dynamically learn a short document graph that facilitates effective label propagation among similar short texts.
arXiv Detail & Related papers (2021-10-30T05:33:05Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z) - Exploiting Contextual Information with Deep Neural Networks [5.787117733071416]
We show that contextual information can be exploited in 2 fundamentally different ways: implicitly and explicitly.
In this thesis, we show that contextual information can be exploited in 2 fundamentally different ways: implicitly and explicitly.
arXiv Detail & Related papers (2020-06-21T03:40:30Z) - Neural Data-to-Text Generation via Jointly Learning the Segmentation and
Correspondence [48.765579605145454]
We propose to explicitly segment target text into fragment units and align them with their data correspondences.
The resulting architecture maintains the same expressive power as neural attention models.
On both E2E and WebNLG benchmarks, we show the proposed model consistently outperforms its neural attention counterparts.
arXiv Detail & Related papers (2020-05-03T14:28:28Z) - Exploiting Structured Knowledge in Text via Graph-Guided Representation
Learning [73.0598186896953]
We present two self-supervised tasks learning over raw text with the guidance from knowledge graphs.
Building upon entity-level masked language models, our first contribution is an entity masking scheme.
In contrast to existing paradigms, our approach uses knowledge graphs implicitly, only during pre-training.
arXiv Detail & Related papers (2020-04-29T14:22:42Z) - CrowdTSC: Crowd-based Neural Networks for Text Sentiment Classification [26.75362694599748]
We propose Crowd-based neural networks for Text Sentiment Classification (CrowdTSC)
We design and post the questions on a crowdsourcing platform to collect the keywords in texts.
We present an attention-based neural network and a hybrid neural network, which incorporate the collected keywords as human beings' guidance.
arXiv Detail & Related papers (2020-04-26T14:08:15Z) - Towards Accurate Scene Text Recognition with Semantic Reasoning Networks [52.86058031919856]
We propose a novel end-to-end trainable framework named semantic reasoning network (SRN) for accurate scene text recognition.
GSRM is introduced to capture global semantic context through multi-way parallel transmission.
Results on 7 public benchmarks, including regular text, irregular text and non-Latin long text, verify the effectiveness and robustness of the proposed method.
arXiv Detail & Related papers (2020-03-27T09:19:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.