Combining word embeddings and convolutional neural networks to detect
duplicated questions
- URL: http://arxiv.org/abs/2006.04513v1
- Date: Mon, 8 Jun 2020 12:30:25 GMT
- Title: Combining word embeddings and convolutional neural networks to detect
duplicated questions
- Authors: Yoan Dimitrov
- Abstract summary: We propose a simple approach to identifying semantically similar questions by combining the strengths of word embeddings and Convolutional Neural Networks.
Our network is trained on the Quora dataset, which contains over 400k question pairs.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Detecting semantic similarities between sentences is still a challenge today
due to the ambiguity of natural languages. In this work, we propose a simple
approach to identifying semantically similar questions by combining the
strengths of word embeddings and Convolutional Neural Networks (CNNs). In
addition, we demonstrate how the cosine similarity metric can be used to
effectively compare feature vectors. Our network is trained on the Quora
dataset, which contains over 400k question pairs. We experiment with different
embedding approaches such as Word2Vec, Fasttext, and Doc2Vec and investigate
the effects these approaches have on model performance. Our model achieves
competitive results on the Quora dataset and complements the well-established
evidence that CNNs can be utilized for paraphrase detection tasks.
Related papers
- Syntax and Semantics Meet in the "Middle": Probing the Syntax-Semantics
Interface of LMs Through Agentivity [68.8204255655161]
We present the semantic notion of agentivity as a case study for probing such interactions.
This suggests LMs may potentially serve as more useful tools for linguistic annotation, theory testing, and discovery.
arXiv Detail & Related papers (2023-05-29T16:24:01Z) - Revealing Similar Semantics Inside CNNs: An Interpretable Concept-based
Comparison of Feature Spaces [0.0]
Safety-critical applications require transparency in artificial intelligence components.
convolutional neural networks (CNNs) widely used for perception tasks lack inherent interpretability.
We propose two methods for estimating the layer-wise similarity between semantic information inside CNN latent spaces.
arXiv Detail & Related papers (2023-04-30T13:53:39Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Explaining Neural Network Predictions on Sentence Pairs via Learning
Word-Group Masks [21.16662651409811]
We propose the Group Mask (GMASK) method to implicitly detect word correlations by grouping correlated words from the input text pair together.
The proposed method is evaluated with two different model architectures (decomposable attention model and BERT) across four datasets.
arXiv Detail & Related papers (2021-04-09T17:14:34Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - On the Effects of Using word2vec Representations in Neural Networks for
Dialogue Act Recognition [0.6767885381740952]
We propose a new deep neural network that explores recurrent models to capture word sequences within sentences.
We validate this model on three languages: English, French and Czech.
arXiv Detail & Related papers (2020-10-22T07:21:17Z) - Sequential Sentence Matching Network for Multi-turn Response Selection
in Retrieval-based Chatbots [45.920841134523286]
We propose a matching network, called sequential sentence matching network (S2M), to use the sentence-level semantic information to address the problem.
Firstly, we find that by using the sentence-level semantic information, the network successfully addresses the problem and gets a significant improvement on matching, resulting in a state-of-the-art performance.
arXiv Detail & Related papers (2020-05-16T09:47:19Z) - Information-Theoretic Probing for Linguistic Structure [74.04862204427944]
We propose an information-theoretic operationalization of probing as estimating mutual information.
We evaluate on a set of ten typologically diverse languages often underrepresented in NLP research.
arXiv Detail & Related papers (2020-04-07T01:06:36Z) - Expressing Objects just like Words: Recurrent Visual Embedding for
Image-Text Matching [102.62343739435289]
Existing image-text matching approaches infer the similarity of an image-text pair by capturing and aggregating the affinities between the text and each independent object of the image.
We propose a Dual Path Recurrent Neural Network (DP-RNN) which processes images and sentences symmetrically by recurrent neural networks (RNN)
Our model achieves the state-of-the-art performance on Flickr30K dataset and competitive performance on MS-COCO dataset.
arXiv Detail & Related papers (2020-02-20T00:51:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.