An Innovative CGL-MHA Model for Sarcasm Sentiment Recognition Using the MindSpore Framework
- URL: http://arxiv.org/abs/2411.01264v1
- Date: Sat, 02 Nov 2024 14:33:47 GMT
- Title: An Innovative CGL-MHA Model for Sarcasm Sentiment Recognition Using the MindSpore Framework
- Authors: Zhenkai Qin, Qining Luo, Xunyi Nong,
- Abstract summary: Sarcasm conveys negative emotions through ostensibly positive or exaggerated language.
We propose an innovative sarcasm detection model integrating CNN, Gated Recurrent Units (GRU), Long Short-Term Memory (LSTM), and Multi-Head Attention mechanisms.
Experiments on two sarcasm detection datasets, Headlines and Riloff, demonstrate that the model achieves an accuracy of 81.20% and an F1 score of 80.77%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The pervasive use of the Internet and social media introduces significant challenges to automated sentiment analysis, particularly for sarcastic expressions in user-generated content. Sarcasm conveys negative emotions through ostensibly positive or exaggerated language, complicating its detection within natural language processing tasks. To address this, we propose an innovative sarcasm detection model integrating Convolutional Neural Networks (CNN), Gated Recurrent Units (GRU), Long Short-Term Memory (LSTM), and Multi-Head Attention mechanisms. The CNN component captures local n-gram features, while GRU and LSTM layers model sequential dependencies and contextual information. Multi-Head Attention enhances the model's focus on relevant parts of the input, improving interpretability. Experiments on two sarcasm detection datasets, Headlines and Riloff, demonstrate that the model achieves an accuracy of 81.20% and an F1 score of 80.77% on Headlines, and an accuracy of 79.72% with an F1 score of 61.39% on Riloff, outperforming traditional models. These results validate the effectiveness of our hybrid approach for sarcasm detection in social media texts.
Related papers
- Opinion Mining and Analysis Using Hybrid Deep Neural Networks [0.0]
We introduce a hybrid deep neural network model that combines a bidirectional gated recurrent unit (BGRU) and long short-term memory (LSTM) layers to improve sentiment analysis.<n>Our model exhibited a noteworthy enhancement in recall for negative sentiments, escalating from 86% (unbalanced dataset) to 96% (balanced dataset)
arXiv Detail & Related papers (2025-11-16T19:04:44Z) - Efficient Hate Speech Detection: Evaluating 38 Models from Traditional Methods to Transformers [0.0]
This study evaluates 38 model configurations in detecting hate speech across datasets ranging from 6.5K to 451K samples.<n>Our results show that transformers, particularly RoBERTa, consistently achieve superior performance with accuracy and F1-scores exceeding 90%.
arXiv Detail & Related papers (2025-09-14T21:17:04Z) - Enhancing Generalization in PPG-Based Emotion Measurement with a CNN-TCN-LSTM Model [5.791020333796703]
This paper introduces a novel hybrid architecture that combines Convolutional Neural Networks (CNNs), Long Short-Term Memory networks (LSTMs), and Temporal Convolutional Networks (TCNs)<n>Our results show that the proposed solution outperforms the state-of-the-art CNN architecture, as well as a CNN-LSTM model, in emotion recognition tasks with PPG signals.
arXiv Detail & Related papers (2025-07-10T16:30:44Z) - SignBart -- New approach with the skeleton sequence for Isolated Sign language Recognition [0.17578923069457017]
This study presents a new novel SLR approach that overcomes the challenge of independently extracting meaningful information from the x and y coordinates of skeleton sequences.<n>With only 749,888 parameters, the model achieves 96.04% accuracy on the LSA-64 dataset.<n>The model also demonstrates excellent performance and generalization across WLASL and ASL-Citizen datasets.
arXiv Detail & Related papers (2025-06-18T07:07:36Z) - A Hybrid Neural Network with Smart Skip Connections for High-Precision, Low-Latency EMG-Based Hand Gesture Recognition [0.2356141385409842]
This paper presents a new hybrid neural network named ConSGruNet for precise and efficient hand gesture recognition.
The proposed model boasts an accuracy of 99.7% in classifying 53 classes in just 25 milliseconds.
arXiv Detail & Related papers (2025-03-12T04:01:32Z) - Representation Learning with Parameterised Quantum Circuits for Advancing Speech Emotion Recognition [37.98283871637917]
Speech Emotion Recognition (SER) is a complex task in human-computer interaction due to the intricate dependencies of features and the overlapping nature of emotional expressions conveyed through speech.
This paper introduces a hybrid classical-quantum framework that integrates volutionised Quantum Circuits with conventional Conal Neural Network (CNN) architectures.
By leveraging quantum properties such as superposition and entanglement, the proposed model enhances feature representation and captures complex dependencies more effectively than classical methods.
arXiv Detail & Related papers (2025-01-21T11:23:38Z) - Leveraging graph neural networks and mobility data for COVID-19 forecasting [37.9506001142702]
COVID-19 pandemic has victimized over 7 million people to date, prompting diverse research efforts.
Spatio-temporal models combining mobility data with machine learning have gained attention for disease forecasting.
Here, we explore Graph Convolutional Recurrent Network (GCRN) and Graph Convolutional Long ShortTerm Memory (GTM)
The aim is to forecast future values of COVID-19 cases in Brazil and China by leveraging human mobility networks.
arXiv Detail & Related papers (2025-01-20T19:52:31Z) - MC-QDSNN: Quantized Deep evolutionary SNN with Multi-Dendritic Compartment Neurons for Stress Detection using Physiological Signals [1.474723404975345]
This work proposes Multi-Compartment Leaky (MCLeaky) neuron as a viable alternative for efficient processing of time series data.
The proposed MCLeaky neuron based Spiking Neural Network model and its quantized variant were benchmarked against state-of-the-art (SOTA) Spiking LSTMs.
Results show that networks with MCLeaky activation neuron managed a superior accuracy of 98.8% to detect stress.
arXiv Detail & Related papers (2024-10-07T12:48:03Z) - Accelerating Domain-Aware Electron Microscopy Analysis Using Deep Learning Models with Synthetic Data and Image-Wide Confidence Scoring [0.0]
We create a physics-based synthetic image and data generator, resulting in a machine learning model that achieves comparable precision (0.86), recall (0.63), F1 scores (0.71), and engineering property predictions (R2=0.82)
Our study demonstrates that synthetic data can eliminate human reliance in ML and provides a means for domain awareness in cases where many feature detections per image are needed.
arXiv Detail & Related papers (2024-08-02T20:15:15Z) - LGB: Language Model and Graph Neural Network-Driven Social Bot Detection [43.92522451274129]
Malicious social bots achieve their malicious purposes by spreading misinformation and inciting social public opinion.
We propose a novel social bot detection framework LGB, which consists of two main components: language model (LM) and graph neural network (GNN)
Experiments on two real-world datasets demonstrate that LGB consistently outperforms state-of-the-art baseline models by up to 10.95%.
arXiv Detail & Related papers (2024-06-13T02:47:38Z) - Non-Intrusive Speech Intelligibility Prediction for Hearing-Impaired
Users using Intermediate ASR Features and Human Memory Models [29.511898279006175]
This work combines the use ofWhisper ASR decoder layer representations as neural network input features with an exemplar-based, psychologically motivated model of human memory to predict human intelligibility ratings for hearing-aid users.
Substantial performance improvement over an established intrusive HASPI baseline system is found, including on enhancement systems and listeners unseen in the training data, with a root mean squared error of 25.3 compared with the baseline of 28.7.
arXiv Detail & Related papers (2024-01-24T17:31:07Z) - Sparse Multitask Learning for Efficient Neural Representation of Motor
Imagery and Execution [30.186917337606477]
We introduce a sparse multitask learning framework for motor imagery (MI) and motor execution (ME) tasks.
Given a dual-task CNN model for MI-ME classification, we apply a saliency-based sparsification approach to prune superfluous connections.
Our results indicate that this tailored sparsity can mitigate the overfitting problem and improve the test performance with small amount of data.
arXiv Detail & Related papers (2023-12-10T09:06:16Z) - Systematic Architectural Design of Scale Transformed Attention Condenser
DNNs via Multi-Scale Class Representational Response Similarity Analysis [93.0013343535411]
We propose a novel type of analysis called Multi-Scale Class Representational Response Similarity Analysis (ClassRepSim)
We show that adding STAC modules to ResNet style architectures can result in up to a 1.6% increase in top-1 accuracy.
Results from ClassRepSim analysis can be used to select an effective parameterization of the STAC module resulting in competitive performance.
arXiv Detail & Related papers (2023-06-16T18:29:26Z) - LDEB -- Label Digitization with Emotion Binarization and Machine
Learning for Emotion Recognition in Conversational Dialogues [0.0]
Emotion recognition in conversations (ERC) is vital to the advancements of conversational AI and its applications.
The conversational dialogues present a unique problem where each dialogue depicts nested emotions that entangle the association between the emotional feature descriptors and emotion type (or label)
We proposed a novel approach called Label Digitization with Emotion Binarization (LDEB) that disentangles the twists by utilizing the text normalization and 7-bit digital encoding techniques.
arXiv Detail & Related papers (2023-06-03T20:37:46Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - HARDC : A novel ECG-based heartbeat classification method to detect
arrhythmia using hierarchical attention based dual structured RNN with
dilated CNN [3.8791511769387625]
We have developed a novel hybrid hierarchical attention-based bidirectional recurrent neural network with dilated CNN (HARDC) method for arrhythmia classification.
The proposed HARDC fully exploits the dilated CNN and bidirectional recurrent neural network unit (BiGRU-BiLSTM) architecture to generate fusion features.
Our results indicate that an automated and highly computed method to classify multiple types of arrhythmia signals holds considerable promise.
arXiv Detail & Related papers (2023-03-06T13:26:29Z) - Light-weighted CNN-Attention based architecture for Hand Gesture
Recognition via ElectroMyography [19.51045409936039]
We propose a light-weighted hybrid architecture (HDCAM) based on Convolutional Neural Network (CNN) and attention mechanism.
The proposed HDCAM model with 58,441 parameters reached a new state-of-the-art (SOTA) performance with 82.91% and 81.28% accuracy on window sizes of 300 ms and 200 ms for classifying 17 hand gestures.
arXiv Detail & Related papers (2022-10-27T02:12:07Z) - Sign Language Recognition via Skeleton-Aware Multi-Model Ensemble [71.97020373520922]
Sign language is commonly used by deaf or mute people to communicate.
We propose a novel Multi-modal Framework with a Global Ensemble Model (GEM) for isolated Sign Language Recognition ( SLR)
Our proposed SAM- SLR-v2 framework is exceedingly effective and achieves state-of-the-art performance with significant margins.
arXiv Detail & Related papers (2021-10-12T16:57:18Z) - Neural Data-to-Text Generation via Jointly Learning the Segmentation and
Correspondence [48.765579605145454]
We propose to explicitly segment target text into fragment units and align them with their data correspondences.
The resulting architecture maintains the same expressive power as neural attention models.
On both E2E and WebNLG benchmarks, we show the proposed model consistently outperforms its neural attention counterparts.
arXiv Detail & Related papers (2020-05-03T14:28:28Z) - Analyzing Redundancy in Pretrained Transformer Models [41.07850306314594]
We define a notion of Redundancy, which we categorize into two classes: General Redundancy and Task-specific Redundancy.
We present an efficient feature-based transfer learning procedure, which maintains 97% performance while using at-most 10% of the original neurons.
arXiv Detail & Related papers (2020-04-08T14:29:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.