A Novel Deep Learning Method for Textual Sentiment Analysis
- URL: http://arxiv.org/abs/2102.11651v1
- Date: Tue, 23 Feb 2021 12:11:36 GMT
- Title: A Novel Deep Learning Method for Textual Sentiment Analysis
- Authors: Hossein Sadr, Mozhdeh Nazari Solimandarabi, Mir Mohsen Pedram,
Mohammad Teshnehlab
- Abstract summary: This paper proposes a convolutional neural network integrated with a hierarchical attention layer to extract informative words.
The proposed model has higher classification accuracy and can extract informative words.
Applying incremental transfer learning can significantly enhance the classification performance.
- Score: 3.0711362702464675
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Sentiment analysis is known as one of the most crucial tasks in the field of
natural language processing and Convolutional Neural Network (CNN) is one of
those prominent models that is commonly used for this aim. Although
convolutional neural networks have obtained remarkable results in recent years,
they are still confronted with some limitations. Firstly, they consider that
all words in a sentence have equal contributions in the sentence meaning
representation and are not able to extract informative words. Secondly, they
require a large number of training data to obtain considerable results while
they have many parameters that must be accurately adjusted. To this end, a
convolutional neural network integrated with a hierarchical attention layer is
proposed which is able to extract informative words and assign them higher
weight. Moreover, the effect of transfer learning that transfers knowledge
learned in the source domain to the target domain with the aim of improving the
performance is also explored. Based on the empirical results, the proposed
model not only has higher classification accuracy and can extract informative
words but also applying incremental transfer learning can significantly enhance
the classification performance.
Related papers
- An Empirical Investigation of Commonsense Self-Supervision with
Knowledge Graphs [67.23285413610243]
Self-supervision based on the information extracted from large knowledge graphs has been shown to improve the generalization of language models.
We study the effect of knowledge sampling strategies and sizes that can be used to generate synthetic data for adapting language models.
arXiv Detail & Related papers (2022-05-21T19:49:04Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Context Decoupling Augmentation for Weakly Supervised Semantic
Segmentation [53.49821324597837]
Weakly supervised semantic segmentation is a challenging problem that has been deeply studied in recent years.
We present a Context Decoupling Augmentation ( CDA) method to change the inherent context in which the objects appear.
To validate the effectiveness of the proposed method, extensive experiments on PASCAL VOC 2012 dataset with several alternative network architectures demonstrate that CDA can boost various popular WSSS methods to the new state-of-the-art by a large margin.
arXiv Detail & Related papers (2021-03-02T15:05:09Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z) - Malicious Network Traffic Detection via Deep Learning: An Information
Theoretic View [0.0]
We study how homeomorphism affects learned representation of a malware traffic dataset.
Our results suggest that although the details of learned representations and the specific coordinate system defined over the manifold of all parameters differ slightly, the functional approximations are the same.
arXiv Detail & Related papers (2020-09-16T15:37:44Z) - Deep Learning based, end-to-end metaphor detection in Greek language
with Recurrent and Convolutional Neural Networks [0.0]
This paper presents and benchmarks a number of end-to-end Deep Learning based models for metaphor detection in Greek.
We combine Convolutional Neural Networks and Recurrent Neural Networks with representation learning to bear on the metaphor detection problem for the Greek language.
arXiv Detail & Related papers (2020-07-23T12:02:40Z) - Adversarially-Trained Deep Nets Transfer Better: Illustration on Image
Classification [53.735029033681435]
Transfer learning is a powerful methodology for adapting pre-trained deep neural networks on image recognition tasks to new domains.
In this work, we demonstrate that adversarially-trained models transfer better than non-adversarially-trained models.
arXiv Detail & Related papers (2020-07-11T22:48:42Z) - An Evaluation of Recent Neural Sequence Tagging Models in Turkish Named
Entity Recognition [5.161531917413708]
We propose a transformer-based network with a conditional random field layer that leads to the state-of-the-art result.
Our study contributes to the literature that quantifies the impact of transfer learning on processing morphologically rich languages.
arXiv Detail & Related papers (2020-05-14T06:54:07Z) - Neurals Networks for Projecting Named Entities from English to Ewondo [6.058868817939519]
We propose a new distributional representation of words to project named entities from a rich language to a low-resource one.
Although the proposed method reached appreciable results, the size of the used neural network was too large.
In this paper, we show experimentally that the same results can be obtained using a smaller neural network.
arXiv Detail & Related papers (2020-03-29T22:05:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.