Evolution of Semantic Similarity -- A Survey
- URL: http://arxiv.org/abs/2004.13820v2
- Date: Sat, 30 Jan 2021 15:57:06 GMT
- Title: Evolution of Semantic Similarity -- A Survey
- Authors: Dhivya Chandrasekaran and Vijay Mago
- Abstract summary: Estimating semantic similarity between text data is a challenging and open research problem in the field of Natural Language Processing (NLP)
Various semantic similarity methods have been proposed over the years to address this issue.
This survey article traces the evolution of such methods, categorizing them based on their underlying principles as knowledge-based, corpus-based, deep neural network-based methods, and hybrid methods.
- Score: 8.873705500708196
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Estimating the semantic similarity between text data is one of the
challenging and open research problems in the field of Natural Language
Processing (NLP). The versatility of natural language makes it difficult to
define rule-based methods for determining semantic similarity measures. In
order to address this issue, various semantic similarity methods have been
proposed over the years. This survey article traces the evolution of such
methods, categorizing them based on their underlying principles as
knowledge-based, corpus-based, deep neural network-based methods, and hybrid
methods. Discussing the strengths and weaknesses of each method, this survey
provides a comprehensive view of existing systems in place, for new researchers
to experiment and develop innovative ideas to address the issue of semantic
similarity.
Related papers
- A study of conceptual language similarity: comparison and evaluation [0.3093890460224435]
An interesting line of research in natural language processing (NLP) aims to incorporate linguistic typology to bridge linguistic diversity.
Recent work has introduced a novel approach to defining language similarity based on how they represent basic concepts.
In this work, we study the conceptual similarity in detail and evaluate it extensively on a binary classification task.
arXiv Detail & Related papers (2023-05-22T18:28:02Z) - A Human Word Association based model for topic detection in social networks [1.8749305679160366]
This paper introduces a topic detection framework for social networks based on the concept of imitating the mental ability of word association.
The performance of this framework is evaluated using the FA-CUP dataset, a benchmark in the field of topic detection.
arXiv Detail & Related papers (2023-01-30T17:10:34Z) - A Survey of Text Representation Methods and Their Genealogy [0.0]
In recent years, with the advent of highly scalable artificial-neural-network-based text representation methods the field of natural language processing has seen unprecedented growth and sophistication.
We provide a survey of current approaches, by arranging them in a genealogy, and by conceptualizing a taxonomy of text representation methods to examine and explain the state-of-the-art.
arXiv Detail & Related papers (2022-11-26T15:22:01Z) - An Inclusive Notion of Text [69.36678873492373]
We argue that clarity on the notion of text is crucial for reproducible and generalizable NLP.
We introduce a two-tier taxonomy of linguistic and non-linguistic elements that are available in textual sources and can be used in NLP modeling.
arXiv Detail & Related papers (2022-11-10T14:26:43Z) - Contextualized language models for semantic change detection: lessons
learned [4.436724861363513]
We present a qualitative analysis of the outputs of contextualized embedding-based methods for detecting diachronic semantic change.
Our findings show that contextualized methods can often predict high change scores for words which are not undergoing any real diachronic semantic shift.
Our conclusion is that pre-trained contextualized language models are prone to confound changes in lexicographic senses and changes in contextual variance.
arXiv Detail & Related papers (2022-08-31T23:35:24Z) - Apport des ontologies pour le calcul de la similarit\'e s\'emantique au
sein d'un syst\`eme de recommandation [0.0]
Measurement of the semantic relatedness or likeness between terms, words, or text data plays an important role in different applications.
We propose and carry on an approach for the calculation of semantic similarity using in the context of a recommender system.
arXiv Detail & Related papers (2022-05-25T07:27:10Z) - Semantic Search for Large Scale Clinical Ontologies [63.71950996116403]
We present a deep learning approach to build a search system for large clinical vocabularies.
We propose a Triplet-BERT model and a method that generates training data based on semantic training data.
The model is evaluated using five real benchmark data sets and the results show that our approach achieves high results on both free text to concept and concept to searching concept vocabularies.
arXiv Detail & Related papers (2022-01-01T05:15:42Z) - A Survey on Deep Semi-supervised Learning [51.26862262550445]
We first present a taxonomy for deep semi-supervised learning that categorizes existing methods.
We then offer a detailed comparison of these methods in terms of the type of losses, contributions, and architecture differences.
arXiv Detail & Related papers (2021-02-28T16:22:58Z) - Learning explanations that are hard to vary [75.30552491694066]
We show that averaging across examples can favor memorization and patchwork' solutions that sew together different strategies.
We then propose and experimentally validate a simple alternative algorithm based on a logical AND.
arXiv Detail & Related papers (2020-09-01T10:17:48Z) - A Survey on Text Classification: From Shallow to Deep Learning [83.47804123133719]
The last decade has seen a surge of research in this area due to the unprecedented success of deep learning.
This paper fills the gap by reviewing the state-of-the-art approaches from 1961 to 2021.
We create a taxonomy for text classification according to the text involved and the models used for feature extraction and classification.
arXiv Detail & Related papers (2020-08-02T00:09:03Z) - There and Back Again: Revisiting Backpropagation Saliency Methods [87.40330595283969]
Saliency methods seek to explain the predictions of a model by producing an importance map across each input sample.
A popular class of such methods is based on backpropagating a signal and analyzing the resulting gradient.
We propose a single framework under which several such methods can be unified.
arXiv Detail & Related papers (2020-04-06T17:58:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.