Recognizing semantic relation in sentence pairs using Tree-RNNs and
Typed dependencies
- URL: http://arxiv.org/abs/2201.04810v1
- Date: Thu, 13 Jan 2022 06:59:27 GMT
- Title: Recognizing semantic relation in sentence pairs using Tree-RNNs and
Typed dependencies
- Authors: Jeena Kleenankandy, K A Abdul Nazeer
- Abstract summary: This work proposes an improvement to Dependency Tree-RNN (DT-RNN) using the grammatical relationship type identified in the dependency parse.
Experiments on semantic relatedness scoring (SRS) and recognizing textual entailment (RTE) in sentence pairs using SICK dataset show encouraging results.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recursive neural networks (Tree-RNNs) based on dependency trees are
ubiquitous in modeling sentence meanings as they effectively capture semantic
relationships between non-neighborhood words. However, recognizing semantically
dissimilar sentences with the same words and syntax is still a challenge to
Tree-RNNs. This work proposes an improvement to Dependency Tree-RNN (DT-RNN)
using the grammatical relationship type identified in the dependency parse. Our
experiments on semantic relatedness scoring (SRS) and recognizing textual
entailment (RTE) in sentence pairs using SICK (Sentence Involving Compositional
Knowledge) dataset show encouraging results. The model achieved a 2%
improvement in classification accuracy for the RTE task over the DT-RNN model.
The results show that Pearson's and Spearman's correlation measures between the
model's predicted similarity scores and human ratings are higher than those of
standard DT-RNNs.
Related papers
- Entity-Aware Self-Attention and Contextualized GCN for Enhanced Relation Extraction in Long Sentences [5.453850739960517]
We propose a novel model, Entity-aware Self-attention Contextualized GCN (ESC-GCN), which efficiently incorporates syntactic structure of input sentences and semantic context of sequences.
Our model achieves encouraging performance as compared to existing dependency-based and sequence-based models.
arXiv Detail & Related papers (2024-09-15T10:50:51Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Can RNNs learn Recursive Nested Subject-Verb Agreements? [4.094098809740732]
Language processing requires the ability to extract nested tree structures.
Recent advances in Recurrent Neural Networks (RNNs) achieve near-human performance in some language tasks.
arXiv Detail & Related papers (2021-01-06T20:47:02Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - Explaining and Improving Model Behavior with k Nearest Neighbor
Representations [107.24850861390196]
We propose using k nearest neighbor representations to identify training examples responsible for a model's predictions.
We show that kNN representations are effective at uncovering learned spurious associations.
Our results indicate that the kNN approach makes the finetuned model more robust to adversarial inputs.
arXiv Detail & Related papers (2020-10-18T16:55:25Z) - High-order Semantic Role Labeling [86.29371274587146]
This paper introduces a high-order graph structure for the neural semantic role labeling model.
It enables the model to explicitly consider not only the isolated predicate-argument pairs but also the interaction between the predicate-argument pairs.
Experimental results on 7 languages of the CoNLL-2009 benchmark show that the high-order structural learning techniques are beneficial to the strong performing SRL models.
arXiv Detail & Related papers (2020-10-09T15:33:54Z) - Building Interpretable Interaction Trees for Deep NLP Models [38.50154540331266]
Six metrics are proposed to analyze properties of interactions between constituents in a sentence.
Our method is used to quantify word interactions encoded inside the BERT, ELMo, LSTM, CNN, and Transformer networks.
arXiv Detail & Related papers (2020-06-29T10:26:50Z) - Combining word embeddings and convolutional neural networks to detect
duplicated questions [0.0]
We propose a simple approach to identifying semantically similar questions by combining the strengths of word embeddings and Convolutional Neural Networks.
Our network is trained on the Quora dataset, which contains over 400k question pairs.
arXiv Detail & Related papers (2020-06-08T12:30:25Z) - Exploiting Syntactic Structure for Better Language Modeling: A Syntactic
Distance Approach [78.77265671634454]
We make use of a multi-task objective, i.e., the models simultaneously predict words as well as ground truth parse trees in a form called "syntactic distances"
Experimental results on the Penn Treebank and Chinese Treebank datasets show that when ground truth parse trees are provided as additional training signals, the model is able to achieve lower perplexity and induce trees with better quality.
arXiv Detail & Related papers (2020-05-12T15:35:00Z) - Linguistically Driven Graph Capsule Network for Visual Question
Reasoning [153.76012414126643]
We propose a hierarchical compositional reasoning model called the "Linguistically driven Graph Capsule Network"
The compositional process is guided by the linguistic parse tree. Specifically, we bind each capsule in the lowest layer to bridge the linguistic embedding of a single word in the original question with visual evidence.
Experiments on the CLEVR dataset, CLEVR compositional generation test, and FigureQA dataset demonstrate the effectiveness and composition generalization ability of our end-to-end model.
arXiv Detail & Related papers (2020-03-23T03:34:25Z) - An enhanced Tree-LSTM architecture for sentence semantic modeling using
typed dependencies [0.0]
Tree-based Long short term memory (LSTM) network has become state-of-the-art for modeling the meaning of language texts.
This paper proposes an enhanced LSTM architecture, called relation gated LSTM, which can model the relationship between two inputs of a sequence.
We also introduce a Tree-LSTM model called Typed Dependency Tree-LSTM that uses the sentence dependency parse structure and the dependency type to embed sentence meaning into a dense vector.
arXiv Detail & Related papers (2020-02-18T18:10:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.