Comparative Analysis of Libraries for the Sentimental Analysis
- URL: http://arxiv.org/abs/2307.14311v1
- Date: Wed, 26 Jul 2023 17:21:53 GMT
- Title: Comparative Analysis of Libraries for the Sentimental Analysis
- Authors: Wendy Ccoya and Edson Pinto
- Abstract summary: This study is main goal is to provide a comparative comparison of libraries using machine learning methods.
Five Python and R libraries NLTK, Textlob Vader, Transformers (GPT and BERT pretrained), and Tidytext will be used in the study to apply sentiment analysis techniques.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study is main goal is to provide a comparative comparison of libraries
using machine learning methods. Experts in natural language processing (NLP)
are becoming more and more interested in sentiment analysis (SA) of text
changes. The objective of employing NLP text analysis techniques is to
recognize and categorize feelings related to twitter users utterances. In this
examination, issues with SA and the libraries utilized are also looked at.
provides a number of cooperative methods to classify emotional polarity. The
Naive Bayes Classifier, Decision Tree Classifier, Maxent Classifier, Sklearn
Classifier, Sklearn Classifier MultinomialNB, and other conjoint learning
algorithms, according to recent research, are very effective. In the project
will use Five Python and R libraries NLTK, TextBlob, Vader, Transformers (GPT
and BERT pretrained), and Tidytext will be used in the study to apply sentiment
analysis techniques. Four machine learning models Tree of Decisions (DT),
Support Vector Machine (SVM), Naive Bayes (NB), and K-Nearest Neighbor (KNN)
will also be used. To evaluate how well libraries for SA operate in the social
network environment, comparative study was also carried out. The measures to
assess the best algorithms in this experiment, which used a single data set for
each method, were precision, recall, and F1 score. We conclude that the BERT
transformer method with an Accuracy: 0.973 is recommended for sentiment
analysis.
Related papers
- Unsupervised Sentiment Analysis of Plastic Surgery Social Media Posts [91.3755431537592]
The massive collection of user posts across social media platforms is primarily untapped for artificial intelligence (AI) use cases.
Natural language processing (NLP) is a subfield of AI that leverages bodies of documents, known as corpora, to train computers in human-like language understanding.
This study demonstrates that the applied results of unsupervised analysis allow a computer to predict either negative, positive, or neutral user sentiment towards plastic surgery.
arXiv Detail & Related papers (2023-07-05T20:16:20Z) - Enhancing Pashto Text Classification using Language Processing
Techniques for Single And Multi-Label Analysis [0.0]
This study aims to establish an automated classification system for Pashto text.
The study achieved an average testing accuracy rate of 94%.
The use of pre-trained language representation models, such as DistilBERT, showed promising results.
arXiv Detail & Related papers (2023-05-04T23:11:31Z) - Pre-trained Embeddings for Entity Resolution: An Experimental Analysis
[Experiment, Analysis & Benchmark] [65.11858854040544]
We perform a thorough experimental analysis of 12 popular language models over 17 established benchmark datasets.
First, we assess their vectorization overhead for converting all input entities into dense embeddings vectors.
Second, we investigate their blocking performance, performing a detailed scalability analysis, and comparing them with the state-of-the-art deep learning-based blocking method.
Third, we conclude with their relative performance for both supervised and unsupervised matching.
arXiv Detail & Related papers (2023-04-24T08:53:54Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Enabling Classifiers to Make Judgements Explicitly Aligned with Human
Values [73.82043713141142]
Many NLP classification tasks, such as sexism/racism detection or toxicity detection, are based on human values.
We introduce a framework for value-aligned classification that performs prediction based on explicitly written human values in the command.
arXiv Detail & Related papers (2022-10-14T09:10:49Z) - InfoCSE: Information-aggregated Contrastive Learning of Sentence
Embeddings [61.77760317554826]
This paper proposes an information-d contrastive learning framework for learning unsupervised sentence embeddings, termed InfoCSE.
We evaluate the proposed InfoCSE on several benchmark datasets w.r.t the semantic text similarity (STS) task.
Experimental results show that InfoCSE outperforms SimCSE by an average Spearman correlation of 2.60% on BERT-base, and 1.77% on BERT-large.
arXiv Detail & Related papers (2022-10-08T15:53:19Z) - A pipeline and comparative study of 12 machine learning models for text
classification [0.0]
Text-based communication is highly favoured as a communication method, especially in business environments.
Many machine learning methods for text classification have been proposed and incorporated into the services of most email providers.
However, optimising text classification algorithms and finding the right tradeoff on their aggressiveness is still a major research problem.
arXiv Detail & Related papers (2022-04-04T23:51:22Z) - Human-in-the-Loop Disinformation Detection: Stance, Sentiment, or
Something Else? [93.91375268580806]
Both politics and pandemics have recently provided ample motivation for the development of machine learning-enabled disinformation (a.k.a. fake news) detection algorithms.
Existing literature has focused primarily on the fully-automated case, but the resulting techniques cannot reliably detect disinformation on the varied topics, sources, and time scales required for military applications.
By leveraging an already-available analyst as a human-in-the-loop, canonical machine learning techniques of sentiment analysis, aspect-based sentiment analysis, and stance detection become plausible methods to use for a partially-automated disinformation detection system.
arXiv Detail & Related papers (2021-11-09T13:30:34Z) - Comparative Analysis of Machine Learning and Deep Learning Algorithms
for Detection of Online Hate Speech [5.543220407902113]
Several attempts have been made to classify hate speech using machine learning but the state-of-the-art models are not robust enough for practical applications.
In this paper, we explored various feature engineering techniques ranging from different embeddings to conventional NLP algorithms.
We conclude that BERT based embeddings give the most useful features for this problem and have the capacity to be made into a practical robust model.
arXiv Detail & Related papers (2021-04-23T04:19:15Z) - SHAP values for Explaining CNN-based Text Classification Models [10.881494765759829]
This paper develops a methodology to compute SHAP values for local explainability of CNN-based text classification models.
The approach is also extended to compute global scores to assess the importance of features.
arXiv Detail & Related papers (2020-08-26T21:28:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.