A Novel BGCapsule Network for Text Classification
- URL: http://arxiv.org/abs/2007.04302v1
- Date: Thu, 2 Jul 2020 06:07:29 GMT
- Title: A Novel BGCapsule Network for Text Classification
- Authors: Akhilesh Kumar Gangwar and Vadlamani Ravi
- Abstract summary: We propose a novel hybrid architecture viz., BGCapsule, which is a Capsule model preceded by an ensemble of Bidirectional Gated Recurrent Units (BiGRU) for several text classification tasks.
BGCapsule achieves better accuracy compared to the existing methods without the help of any external linguistic knowledge.
- Score: 5.010425616264462
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Several text classification tasks such as sentiment analysis, news
categorization, multi-label classification and opinion classification are
challenging problems even for modern deep learning networks. Recently, Capsule
Networks (CapsNets) are proposed for image classification. It has been shown
that CapsNets have several advantages over Convolutional Neural Networks
(CNNs), while their validity in the domain of text has been less explored. In
this paper, we propose a novel hybrid architecture viz., BGCapsule, which is a
Capsule model preceded by an ensemble of Bidirectional Gated Recurrent Units
(BiGRU) for several text classification tasks. We employed an ensemble of
Bidirectional GRUs for feature extraction layer preceding the primary capsule
layer. The hybrid architecture, after performing basic pre-processing steps,
consists of five layers: an embedding layer based on GloVe, a BiGRU based
ensemble layer, a primary capsule layer, a flatten layer and fully connected
ReLU layer followed by a fully connected softmax layer. In order to evaluate
the effectiveness of BGCapsule, we conducted extensive experiments on five
benchmark datasets (ranging from 10,000 records to 700,000 records) including
Movie Review (MR Imdb 2005), AG News dataset, Dbpedia ontology dataset, Yelp
Review Full dataset and Yelp review polarity dataset. These benchmarks cover
several text classification tasks such as news categorization, sentiment
analysis, multiclass classification, multi-label classification and opinion
classification. We found that our proposed architecture (BGCapsule) achieves
better accuracy compared to the existing methods without the help of any
external linguistic knowledge such as positive sentiment keywords and negative
sentiment keywords. Further, BGCapsule converged faster compared to other
extant techniques.
Related papers
- UniCell: Universal Cell Nucleus Classification via Prompt Learning [76.11864242047074]
We propose a universal cell nucleus classification framework (UniCell)
It employs a novel prompt learning mechanism to uniformly predict the corresponding categories of pathological images from different dataset domains.
In particular, our framework adopts an end-to-end architecture for nuclei detection and classification, and utilizes flexible prediction heads for adapting various datasets.
arXiv Detail & Related papers (2024-02-20T11:50:27Z) - Learning the Right Layers: a Data-Driven Layer-Aggregation Strategy for
Semi-Supervised Learning on Multilayer Graphs [2.752817022620644]
Clustering (or community detection) on multilayer graphs poses several additional complications.
One of the major challenges is to establish the extent to which each layer contributes to the cluster iteration assignment.
We propose a parameter-free Laplacian-regularized model that learns an optimal nonlinear combination of the different layers from the available input labels.
arXiv Detail & Related papers (2023-05-31T19:50:11Z) - WLD-Reg: A Data-dependent Within-layer Diversity Regularizer [98.78384185493624]
Neural networks are composed of multiple layers arranged in a hierarchical structure jointly trained with a gradient-based optimization.
We propose to complement this traditional 'between-layer' feedback with additional 'within-layer' feedback to encourage the diversity of the activations within the same layer.
We present an extensive empirical study confirming that the proposed approach enhances the performance of several state-of-the-art neural network models in multiple tasks.
arXiv Detail & Related papers (2023-01-03T20:57:22Z) - Semantic Guided Level-Category Hybrid Prediction Network for
Hierarchical Image Classification [8.456482280676884]
Hierarchical classification (HC) assigns each object with multiple labels organized into a hierarchical structure.
We propose a novel semantic guided level-category hybrid prediction network (SGLCHPN) that can jointly perform the level and category prediction in an end-to-end manner.
arXiv Detail & Related papers (2022-11-22T13:49:10Z) - A Capsule Network for Hierarchical Multi-Label Image Classification [2.507647327384289]
Hierarchical multi-label classification applies when a multi-class image classification problem is arranged into smaller ones based upon a hierarchy or taxonomy.
We propose a multi-label capsule network (ML-CapsNet) for hierarchical classification.
arXiv Detail & Related papers (2022-09-13T04:17:08Z) - Dual-State Capsule Networks for Text Classification [2.0305676256390934]
This paper presents a novel Dual-State Capsule (DS-Caps) network-based technique for text classification.
Two varieties of states, namely sentence-level and word-level, are integrated with capsule layers to capture deeper context-level information.
The DS-Caps networks outperform the existing capsule network architectures for multiple datasets.
arXiv Detail & Related papers (2021-09-10T09:59:55Z) - An evidential classifier based on Dempster-Shafer theory and deep
learning [6.230751621285322]
We propose a new classification system based on Dempster-Shafer (DS) theory and a convolutional neural network (CNN) architecture for set-valued classification.
Experiments on image recognition, signal processing, and semantic-relationship classification tasks demonstrate that the proposed combination of deep CNN, DS layer, and expected utility layer makes it possible to improve classification accuracy.
arXiv Detail & Related papers (2021-03-25T01:29:05Z) - Minimally-Supervised Structure-Rich Text Categorization via Learning on
Text-Rich Networks [61.23408995934415]
We propose a novel framework for minimally supervised categorization by learning from the text-rich network.
Specifically, we jointly train two modules with different inductive biases -- a text analysis module for text understanding and a network learning module for class-discriminative, scalable network learning.
Our experiments show that given only three seed documents per category, our framework can achieve an accuracy of about 92%.
arXiv Detail & Related papers (2021-02-23T04:14:34Z) - Hierarchical Graph Capsule Network [78.4325268572233]
We propose hierarchical graph capsule network (HGCN) that can jointly learn node embeddings and extract graph hierarchies.
To learn the hierarchical representation, HGCN characterizes the part-whole relationship between lower-level capsules (part) and higher-level capsules (whole)
arXiv Detail & Related papers (2020-12-16T04:13:26Z) - Graph Cross Networks with Vertex Infomax Pooling [69.38969610952927]
We propose a novel graph cross network (GXN) to achieve comprehensive feature learning from multiple scales of a graph.
Based on trainable hierarchical representations of a graph, GXN enables the interchange of intermediate features across scales to promote information flow.
arXiv Detail & Related papers (2020-10-05T06:34:23Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.