Text Classification with Lexicon from PreAttention Mechanism
- URL: http://arxiv.org/abs/2002.07591v1
- Date: Tue, 18 Feb 2020 14:40:20 GMT
- Title: Text Classification with Lexicon from PreAttention Mechanism
- Authors: QingBiao LI (Beijing University of Posts and Telecommunications),
Chunhua Wu (Beijing University of Posts and Telecommunications), Kangfeng
Zheng (Beijing University of Posts and Telecommunications)
- Abstract summary: A comprehensive and high-quality lexicon plays a crucial role in traditional text classification approaches.
We propose a Pre-Attention mechanism for text classification, which can learn attention of different words according to their effects in the classification tasks.
We get 90.5% accuracy on Stanford Large Movie Review dataset, 82.3% on Subjectivity dataset, 93.7% on Movie Reviews.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A comprehensive and high-quality lexicon plays a crucial role in traditional
text classification approaches. And it improves the utilization of the
linguistic knowledge. Although it is helpful for the task, the lexicon has got
little attention in recent neural network models. Firstly, getting a
high-quality lexicon is not easy. We lack an effective automated lexicon
extraction method, and most lexicons are hand crafted, which is very
inefficient for big data. What's more, there is no an effective way to use a
lexicon in a neural network. To address those limitations, we propose a
Pre-Attention mechanism for text classification in this paper, which can learn
attention of different words according to their effects in the classification
tasks. The words with different attention can form a domain lexicon.
Experiments on three benchmark text classification tasks show that our models
get competitive result comparing with the state-of-the-art methods. We get
90.5% accuracy on Stanford Large Movie Review dataset, 82.3% on Subjectivity
dataset, 93.7% on Movie Reviews. And compared with the text classification
model without Pre-Attention mechanism, those with Pre-Attention mechanism
improve by 0.9%-2.4% accuracy, which proves the validity of the Pre-Attention
mechanism. In addition, the Pre-Attention mechanism performs well followed by
different types of neural networks (e.g., convolutional neural networks and
Long Short-Term Memory networks). For the same dataset, when we use
Pre-Attention mechanism to get attention value followed by different neural
networks, those words with high attention values have a high degree of
coincidence, which proves the versatility and portability of the Pre-Attention
mechanism. we can get stable lexicons by attention values, which is an
inspiring method of information extraction.
Related papers
- Text Classification Based on Knowledge Graphs and Improved Attention
Mechanism [12.008192698720947]
The model operates at both character and word levels to deepen its understanding by integrating the concepts.
Its performance is demonstrated on datasets such as AGNews, Ohsumed, and TagMyNews, achieving accuracy of 75.1%, 58.7%, and 68.5% respectively.
arXiv Detail & Related papers (2024-01-07T22:20:55Z) - Multi-label Text Classification using GloVe and Neural Network Models [0.27195102129094995]
Existing solutions include traditional machine learning and deep neural networks for predictions.
This paper proposes a method utilizing the bag-of-words model approach based on the GloVe model and the CNN-BiLSTM network.
The method achieves an accuracy rate of 87.26% on the test set and an F1 score of 0.8737, showcasing promising results.
arXiv Detail & Related papers (2023-10-25T01:30:26Z) - Three ways to improve feature alignment for open vocabulary detection [88.65076922242184]
Key problem in zero-shot open vocabulary detection is how to align visual and text features, so that the detector performs well on unseen classes.
Previous approaches train the feature pyramid and detection head from scratch, which breaks the vision-text feature alignment established during pretraining.
We propose three methods to alleviate these issues. Firstly, a simple scheme is used to augment the text embeddings which prevents overfitting to a small number of classes seen during training.
Secondly, the feature pyramid network and the detection head are modified to include trainable shortcuts.
Finally, a self-training approach is used to leverage a larger corpus of
arXiv Detail & Related papers (2023-03-23T17:59:53Z) - Open Vocabulary Electroencephalography-To-Text Decoding and Zero-shot
Sentiment Classification [78.120927891455]
State-of-the-art brain-to-text systems have achieved great success in decoding language directly from brain signals using neural networks.
In this paper, we extend the problem to open vocabulary Electroencephalography(EEG)-To-Text Sequence-To-Sequence decoding and zero-shot sentence sentiment classification on natural reading tasks.
Our model achieves a 40.1% BLEU-1 score on EEG-To-Text decoding and a 55.6% F1 score on zero-shot EEG-based ternary sentiment classification, which significantly outperforms supervised baselines.
arXiv Detail & Related papers (2021-12-05T21:57:22Z) - Speaker Embedding-aware Neural Diarization for Flexible Number of
Speakers with Textual Information [55.75018546938499]
We propose the speaker embedding-aware neural diarization (SEND) method, which predicts the power set encoded labels.
Our method achieves lower diarization error rate than the target-speaker voice activity detection.
arXiv Detail & Related papers (2021-11-28T12:51:04Z) - Intent Classification Using Pre-Trained Embeddings For Low Resource
Languages [67.40810139354028]
Building Spoken Language Understanding systems that do not rely on language specific Automatic Speech Recognition is an important yet less explored problem in language processing.
We present a comparative study aimed at employing a pre-trained acoustic model to perform Spoken Language Understanding in low resource scenarios.
We perform experiments across three different languages: English, Sinhala, and Tamil each with different data sizes to simulate high, medium, and low resource scenarios.
arXiv Detail & Related papers (2021-10-18T13:06:59Z) - Preliminary study on using vector quantization latent spaces for TTS/VC
systems with consistent performance [55.10864476206503]
We investigate the use of quantized vectors to model the latent linguistic embedding.
By enforcing different policies over the latent spaces in the training, we are able to obtain a latent linguistic embedding.
Our experiments show that the voice cloning system built with vector quantization has only a small degradation in terms of perceptive evaluations.
arXiv Detail & Related papers (2021-06-25T07:51:35Z) - Detecting of a Patient's Condition From Clinical Narratives Using
Natural Language Representation [0.3149883354098941]
This paper proposes a joint clinical natural language representation learning and supervised classification framework.
The novel framework jointly discovers distributional syntactic and latent semantic (representation learning) from contextual clinical narrative inputs.
The proposed framework yields an overall classification performance with accuracy, recall, and precision of 89 % and 88 %, 89 %, respectively.
arXiv Detail & Related papers (2021-04-08T17:16:04Z) - A Novel Deep Learning Method for Textual Sentiment Analysis [3.0711362702464675]
This paper proposes a convolutional neural network integrated with a hierarchical attention layer to extract informative words.
The proposed model has higher classification accuracy and can extract informative words.
Applying incremental transfer learning can significantly enhance the classification performance.
arXiv Detail & Related papers (2021-02-23T12:11:36Z) - Does a Hybrid Neural Network based Feature Selection Model Improve Text
Classification? [9.23545668304066]
We propose a hybrid feature selection method for obtaining relevant features.
We then present three ways of implementing a feature selection and neural network pipeline.
We also observed a slight increase in accuracy on some datasets.
arXiv Detail & Related papers (2021-01-22T09:12:19Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.