KETM:A Knowledge-Enhanced Text Matching method
- URL: http://arxiv.org/abs/2308.06235v1
- Date: Fri, 11 Aug 2023 17:08:14 GMT
- Title: KETM:A Knowledge-Enhanced Text Matching method
- Authors: Kexin Jiang, Yahui Zhao, Guozhe Jin, Zhenguo Zhang and Rongyi Cui
- Abstract summary: We introduce a new model for text matching called the Knowledge Enhanced Text Matching model (KETM)
We use Wiktionary to retrieve the text word definitions as our external knowledge.
We fuse text and knowledge using a gating mechanism to learn the ratio of text and knowledge fusion.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text matching is the task of matching two texts and determining the
relationship between them, which has extensive applications in natural language
processing tasks such as reading comprehension, and Question-Answering systems.
The mainstream approach is to compute text representations or to interact with
the text through attention mechanism, which is effective in text matching
tasks. However, the performance of these models is insufficient for texts that
require commonsense knowledge-based reasoning. To this end, in this paper, We
introduce a new model for text matching called the Knowledge Enhanced Text
Matching model (KETM), to enrich contextual representations with real-world
common-sense knowledge from external knowledge sources to enhance our model
understanding and reasoning. First, we use Wiktionary to retrieve the text word
definitions as our external knowledge. Secondly, we feed text and knowledge to
the text matching module to extract their feature vectors. The text matching
module is used as an interaction module by integrating the encoder layer, the
co-attention layer, and the aggregation layer. Specifically, the interaction
process is iterated several times to obtain in-depth interaction information
and extract the feature vectors of text and knowledge by multi-angle pooling.
Then, we fuse text and knowledge using a gating mechanism to learn the ratio of
text and knowledge fusion by a neural network that prevents noise generated by
knowledge. After that, experimental validation on four datasets are carried
out, and the experimental results show that our proposed model performs well on
all four datasets, and the performance of our method is improved compared to
the base model without adding external knowledge, which validates the
effectiveness of our proposed method. The code is available at
https://github.com/1094701018/KETM
Related papers
- Learning Robust Named Entity Recognizers From Noisy Data With Retrieval Augmentation [67.89838237013078]
Named entity recognition (NER) models often struggle with noisy inputs.
We propose a more realistic setting in which only noisy text and its NER labels are available.
We employ a multi-view training framework that improves robust NER without retrieving text during inference.
arXiv Detail & Related papers (2024-07-26T07:30:41Z) - TextFormer: A Query-based End-to-End Text Spotter with Mixed Supervision [61.186488081379]
We propose TextFormer, a query-based end-to-end text spotter with Transformer architecture.
TextFormer builds upon an image encoder and a text decoder to learn a joint semantic understanding for multi-task modeling.
It allows for mutual training and optimization of classification, segmentation, and recognition branches, resulting in deeper feature sharing.
arXiv Detail & Related papers (2023-06-06T03:37:41Z) - TRIE++: Towards End-to-End Information Extraction from Visually Rich
Documents [51.744527199305445]
This paper proposes a unified end-to-end information extraction framework from visually rich documents.
Text reading and information extraction can reinforce each other via a well-designed multi-modal context block.
The framework can be trained in an end-to-end trainable manner, achieving global optimization.
arXiv Detail & Related papers (2022-07-14T08:52:07Z) - DEIM: An effective deep encoding and interaction model for sentence
matching [0.0]
We propose a sentence matching method based on deep encoding and interaction to extract deep semantic information.
In the encoder layer,we refer to the information of another sentence in the process of encoding a single sentence, and later use a algorithm to fuse the information.
In the interaction layer, we use a bidirectional attention mechanism and a self-attention mechanism to obtain deep semantic information.
arXiv Detail & Related papers (2022-03-20T07:59:42Z) - Explaining Neural Network Predictions on Sentence Pairs via Learning
Word-Group Masks [21.16662651409811]
We propose the Group Mask (GMASK) method to implicitly detect word correlations by grouping correlated words from the input text pair together.
The proposed method is evaluated with two different model architectures (decomposable attention model and BERT) across four datasets.
arXiv Detail & Related papers (2021-04-09T17:14:34Z) - Improving Machine Reading Comprehension with Contextualized Commonsense
Knowledge [62.46091695615262]
We aim to extract commonsense knowledge to improve machine reading comprehension.
We propose to represent relations implicitly by situating structured knowledge in a context.
We employ a teacher-student paradigm to inject multiple types of contextualized knowledge into a student machine reader.
arXiv Detail & Related papers (2020-09-12T17:20:01Z) - TRIE: End-to-End Text Reading and Information Extraction for Document
Understanding [56.1416883796342]
We propose a unified end-to-end text reading and information extraction network.
multimodal visual and textual features of text reading are fused for information extraction.
Our proposed method significantly outperforms the state-of-the-art methods in both efficiency and accuracy.
arXiv Detail & Related papers (2020-05-27T01:47:26Z) - Exploiting Structured Knowledge in Text via Graph-Guided Representation
Learning [73.0598186896953]
We present two self-supervised tasks learning over raw text with the guidance from knowledge graphs.
Building upon entity-level masked language models, our first contribution is an entity masking scheme.
In contrast to existing paradigms, our approach uses knowledge graphs implicitly, only during pre-training.
arXiv Detail & Related papers (2020-04-29T14:22:42Z) - Generating Hierarchical Explanations on Text Classification via Feature
Interaction Detection [21.02924712220406]
We build hierarchical explanations by detecting feature interactions.
Such explanations visualize how words and phrases are combined at different levels of the hierarchy.
Experiments show the effectiveness of the proposed method in providing explanations both faithful to models and interpretable to humans.
arXiv Detail & Related papers (2020-04-04T20:56:37Z) - Matching Text with Deep Mutual Information Estimation [0.0]
We present a neural approach for general-purpose text matching with deep mutual information estimation incorporated.
Our approach, Text matching with Deep Info Max (TIM), is integrated with a procedure of unsupervised learning of representations.
We evaluate our text matching approach on several tasks including natural language inference, paraphrase identification, and answer selection.
arXiv Detail & Related papers (2020-03-09T15:25:37Z) - Short Text Classification via Knowledge powered Attention with
Similarity Matrix based CNN [6.6723692875904375]
We propose a knowledge powered attention with similarity matrix based convolutional neural network (KASM) model.
We use knowledge graph (KG) to enrich the semantic representation of short text, specially, the information of parent-entity is introduced in our model.
For the purpose of measuring the importance of knowledge, we introduce the attention mechanisms to choose the important information.
arXiv Detail & Related papers (2020-02-09T12:08:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.