An Insight into The Intricacies of Lingual Paraphrasing Pragmatic
Discourse on The Purpose of Synonyms
- URL: http://arxiv.org/abs/2206.02983v1
- Date: Tue, 7 Jun 2022 02:57:27 GMT
- Title: An Insight into The Intricacies of Lingual Paraphrasing Pragmatic
Discourse on The Purpose of Synonyms
- Authors: Jabir Al Nahian, Abu Kaisar Mohammad Masum, Muntaser Mansur Syed,
Sheikh Abujar
- Abstract summary: We develop an algorithm to paraphrase any text document or paragraphs using WordNet and Natural Language Tool Kit (NLTK)
For 250 paragraphs, our algorithm achieved a paraphrase accuracy of 94.8%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The term "paraphrasing" refers to the process of presenting the sense of an
input text in a new way while preserving fluency. Scientific research
distribution is gaining traction, allowing both rookie and experienced
scientists to participate in their respective fields. As a result, there is now
a massive demand for paraphrase tools that may efficiently and effectively
assist scientists in modifying statements in order to avoid plagiarism. Natural
Language Processing (NLP) is very much important in the realm of the process of
document paraphrasing. We analyze and discuss existing studies on paraphrasing
in the English language in this paper. Finally, we develop an algorithm to
paraphrase any text document or paragraphs using WordNet and Natural Language
Tool Kit (NLTK) and maintain "Using Synonyms" techniques to achieve our result.
For 250 paragraphs, our algorithm achieved a paraphrase accuracy of 94.8%
Related papers
- The Nature of NLP: Analyzing Contributions in NLP Papers [77.31665252336157]
We quantitatively investigate what constitutes NLP research by examining research papers.
Our findings reveal a rising involvement of machine learning in NLP since the early nineties.
In post-2020, there has been a resurgence of focus on language and people.
arXiv Detail & Related papers (2024-09-29T01:29:28Z) - Towards Effective Paraphrasing for Information Disguise [13.356934367660811]
Research on Information Disguise (ID) becomes important when authors' written online communication pertains to sensitive domains.
We propose a framework where, for a given sentence from an author's post, we perform iterative perturbation on the sentence in the direction of paraphrasing.
Our work introduces a novel method of phrase-importance rankings using perplexity scores and involves multi-level phrase substitutions via beam search.
arXiv Detail & Related papers (2023-11-08T21:12:59Z) - An Inclusive Notion of Text [69.36678873492373]
We argue that clarity on the notion of text is crucial for reproducible and generalizable NLP.
We introduce a two-tier taxonomy of linguistic and non-linguistic elements that are available in textual sources and can be used in NLP modeling.
arXiv Detail & Related papers (2022-11-10T14:26:43Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - A Survey On Neural Word Embeddings [0.4822598110892847]
The study of meaning in natural language processing relies on the distributional hypothesis.
The revolutionary idea of distributed representation for a concept is close to the working of a human mind.
Neural word embeddings transformed the whole field of NLP by introducing substantial improvements in all NLP tasks.
arXiv Detail & Related papers (2021-10-05T03:37:57Z) - An Empirical Study on Leveraging Position Embeddings for Target-oriented
Opinion Words Extraction [13.765146062545048]
Target-oriented opinion words extraction (TOWE) is a new subtask of target-oriented sentiment analysis.
We show that BiLSTM-based models can effectively encode position information into word representations.
We also adapt a graph convolutional network (GCN) to enhance word representations by incorporating syntactic information.
arXiv Detail & Related papers (2021-09-02T22:49:45Z) - CitationIE: Leveraging the Citation Graph for Scientific Information
Extraction [89.33938657493765]
We use the citation graph of referential links between citing and cited papers.
We observe a sizable improvement in end-to-end information extraction over the state-of-the-art.
arXiv Detail & Related papers (2021-06-03T03:00:12Z) - Word-level Human Interpretable Scoring Mechanism for Novel Text
Detection Using Tsetlin Machines [16.457778420360537]
We propose a Tsetlin machine architecture for scoring individual words according to their contribution to novelty.
Our approach encodes a description of the novel documents using the linguistic patterns captured by TM clauses.
We then adopt this description to measure how much a word contributes to making documents novel.
arXiv Detail & Related papers (2021-05-10T23:41:14Z) - Semantic Analysis for Automated Evaluation of the Potential Impact of
Research Articles [62.997667081978825]
This paper presents a novel method for vector representation of text meaning based on information theory.
We show how this informational semantics is used for text classification on the basis of the Leicester Scientific Corpus.
We show that an informational approach to representing the meaning of a text has offered a way to effectively predict the scientific impact of research papers.
arXiv Detail & Related papers (2021-04-26T20:37:13Z) - Intrinsic Probing through Dimension Selection [69.52439198455438]
Most modern NLP systems make use of pre-trained contextual representations that attain astonishingly high performance on a variety of tasks.
Such high performance should not be possible unless some form of linguistic structure inheres in these representations, and a wealth of research has sprung up on probing for it.
In this paper, we draw a distinction between intrinsic probing, which examines how linguistic information is structured within a representation, and the extrinsic probing popular in prior work, which only argues for the presence of such information by showing that it can be successfully extracted.
arXiv Detail & Related papers (2020-10-06T15:21:08Z) - An Analysis of Word2Vec for the Italian Language [0.0]
Word representation is fundamental in NLP tasks, because it is precisely from the coding of semantic closeness between words that it is possible to think of teaching a machine to understand text.
In this work, analysing the semantic capacity of the Word2Vec algorithm, an embedding for the Italian language is produced.
arXiv Detail & Related papers (2020-01-25T15:12:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.