Bidirectional Encoder Representations from Transformers (BERT): A
sentiment analysis odyssey
- URL: http://arxiv.org/abs/2007.01127v1
- Date: Thu, 2 Jul 2020 14:23:57 GMT
- Title: Bidirectional Encoder Representations from Transformers (BERT): A
sentiment analysis odyssey
- Authors: Shivaji Alaparthi (Data Scientist, CenturyLink, Bengaluru, India) and
Manit Mishra (Associate Professor, International Management Institute
Bhubaneswar, India)
- Abstract summary: The study puts forth two key insights: (1) relative efficacy of four highly advanced and widely used sentiment analysis techniques; and (2) undisputed superiority of pre-trained advanced supervised deep learning BERT model in sentiment analysis from text data.
We use publicly available labeled corpora of 50,000 movie reviews originally posted on internet movie database (IMDB) for analysis using Sent WordNet lexicon, logistic regression, LSTM, and BERT.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The purpose of the study is to investigate the relative effectiveness of four
different sentiment analysis techniques: (1) unsupervised lexicon-based model
using Sent WordNet; (2) traditional supervised machine learning model using
logistic regression; (3) supervised deep learning model using Long Short-Term
Memory (LSTM); and, (4) advanced supervised deep learning models using
Bidirectional Encoder Representations from Transformers (BERT). We use publicly
available labeled corpora of 50,000 movie reviews originally posted on internet
movie database (IMDB) for analysis using Sent WordNet lexicon, logistic
regression, LSTM, and BERT. The first three models were run on CPU based system
whereas BERT was run on GPU based system. The sentiment classification
performance was evaluated based on accuracy, precision, recall, and F1 score.
The study puts forth two key insights: (1) relative efficacy of four highly
advanced and widely used sentiment analysis techniques; (2) undisputed
superiority of pre-trained advanced supervised deep learning BERT model in
sentiment analysis from text data. This study provides professionals in
analytics industry and academicians working on text analysis key insight
regarding comparative classification performance evaluation of key sentiment
analysis techniques, including the recently developed BERT. This is the first
research endeavor to compare the advanced pre-trained supervised deep learning
model of BERT vis-\`a-vis other sentiment analysis models of LSTM, logistic
regression, and Sent WordNet.
Related papers
- The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - Extensive Evaluation of Transformer-based Architectures for Adverse Drug
Events Extraction [6.78974856327994]
Adverse Event (ADE) extraction is one of the core tasks in digital pharmacovigilance.
We evaluate 19 Transformer-based models for ADE extraction on informal texts.
At the end of our analyses, we identify a list of take-home messages that can be derived from the experimental data.
arXiv Detail & Related papers (2023-06-08T15:25:24Z) - Pre-trained Embeddings for Entity Resolution: An Experimental Analysis
[Experiment, Analysis & Benchmark] [65.11858854040544]
We perform a thorough experimental analysis of 12 popular language models over 17 established benchmark datasets.
First, we assess their vectorization overhead for converting all input entities into dense embeddings vectors.
Second, we investigate their blocking performance, performing a detailed scalability analysis, and comparing them with the state-of-the-art deep learning-based blocking method.
Third, we conclude with their relative performance for both supervised and unsupervised matching.
arXiv Detail & Related papers (2023-04-24T08:53:54Z) - Transformer-based approaches to Sentiment Detection [55.41644538483948]
We examined the performance of four different types of state-of-the-art transformer models for text classification.
The RoBERTa transformer model performs best on the test dataset with a score of 82.6% and is highly recommended for quality predictions.
arXiv Detail & Related papers (2023-03-13T17:12:03Z) - Pre-trained Language Models for Keyphrase Generation: A Thorough
Empirical Study [76.52997424694767]
We present an in-depth empirical study of keyphrase extraction and keyphrase generation using pre-trained language models.
We show that PLMs have competitive high-resource performance and state-of-the-art low-resource performance.
Further results show that in-domain BERT-like PLMs can be used to build strong and data-efficient keyphrase generation models.
arXiv Detail & Related papers (2022-12-20T13:20:21Z) - BERT-Based Combination of Convolutional and Recurrent Neural Network for
Indonesian Sentiment Analysis [0.0]
This research extends the previous hybrid deep learning using BERT representation for Indonesian sentiment analysis.
Our simulation shows that the BERT representation improves the accuracies of all hybrid architectures.
arXiv Detail & Related papers (2022-11-10T00:32:40Z) - Aspect-Based Sentiment Analysis using Local Context Focus Mechanism with
DeBERTa [23.00810941211685]
Aspect-Based Sentiment Analysis (ABSA) is a fine-grained task in the field of sentiment analysis.
Recent DeBERTa model (Decoding-enhanced BERT with disentangled attention) to solve Aspect-Based Sentiment Analysis problem.
arXiv Detail & Related papers (2022-07-06T03:50:31Z) - BERT based sentiment analysis: A software engineering perspective [0.9176056742068814]
The paper presents three different strategies to analyse BERT based model for sentiment analysis.
The experimental results show that the BERT based ensemble approach and the compressed BERT model attain improvements by 6-12% over prevailing tools for the F1 measure on all three datasets.
arXiv Detail & Related papers (2021-06-04T16:28:26Z) - A comprehensive comparative evaluation and analysis of Distributional
Semantic Models [61.41800660636555]
We perform a comprehensive evaluation of type distributional vectors, either produced by static DSMs or obtained by averaging the contextualized vectors generated by BERT.
The results show that the alleged superiority of predict based models is more apparent than real, and surely not ubiquitous.
We borrow from cognitive neuroscience the methodology of Representational Similarity Analysis (RSA) to inspect the semantic spaces generated by distributional models.
arXiv Detail & Related papers (2021-05-20T15:18:06Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.