Email Spam Detection Using Hierarchical Attention Hybrid Deep Learning
Method
- URL: http://arxiv.org/abs/2204.07390v1
- Date: Fri, 15 Apr 2022 09:02:36 GMT
- Title: Email Spam Detection Using Hierarchical Attention Hybrid Deep Learning
Method
- Authors: Sultan Zavrak and Seyhmus Yilmaz
- Abstract summary: This article proposes a novel technique for email spam detection based on a combination of convolutional neural networks, recurrent units, and attention gated mechanisms.
The proposed technique's findings are compared to those of state-of-the-art models and show that our approach outperforms them.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Email is one of the most widely used ways to communicate, with millions of
people and businesses relying on it to communicate and share knowledge and
information on a daily basis. Nevertheless, the rise in email users has
occurred a dramatic increase in spam emails in recent years. Processing and
managing emails properly for individuals and companies are getting increasingly
difficult. This article proposes a novel technique for email spam detection
that is based on a combination of convolutional neural networks, gated
recurrent units, and attention mechanisms. During system training, the network
is selectively focused on necessary parts of the email text. The usage of
convolution layers to extract more meaningful, abstract, and generalizable
features by hierarchical representation is the major contribution of this
study. Additionally, this contribution incorporates cross-dataset evaluation,
which enables the generation of more independent performance results from the
model's training dataset. According to cross-dataset evaluation results, the
proposed technique advances the results of the present attention-based
techniques by utilizing temporal convolutions, which give us more flexible
receptive field sizes are utilized. The suggested technique's findings are
compared to those of state-of-the-art models and show that our approach
outperforms them.
Related papers
- Communication-Efficient Large-Scale Distributed Deep Learning: A Comprehensive Survey [43.57122822150023]
This article surveys the literature on algorithms and technologies aimed at achieving efficient communication in large-scale distributed deep learning.
We first introduce efficient algorithms for model synchronization and communication data compression in the context of large-scale distributed training.
Next, we introduce efficient strategies related to resource allocation and task scheduling for use in distributed training and inference.
arXiv Detail & Related papers (2024-04-09T08:35:04Z) - Data Augmentations in Deep Weight Spaces [89.45272760013928]
We introduce a novel augmentation scheme based on the Mixup method.
We evaluate the performance of these techniques on existing benchmarks as well as new benchmarks we generate.
arXiv Detail & Related papers (2023-11-15T10:43:13Z) - PROMINET: Prototype-based Multi-View Network for Interpretable Email
Response Prediction [12.727146945870809]
This study proposes a Prototype-based Multi-view Network (PROMINET) that incorporates semantic and structural information from email data.
The model maps learned semantic and structural exemplars to observed samples in the training data at different levels of granularity, such as document, sentence, or phrase.
The learned prototypes also show potential for generating suggestions to enhance email text editing and improve the likelihood of effective email responses.
arXiv Detail & Related papers (2023-10-25T16:39:00Z) - Reinforcement Learning Based Multi-modal Feature Fusion Network for
Novel Class Discovery [47.28191501836041]
In this paper, we employ a Reinforcement Learning framework to simulate the cognitive processes of humans.
We also deploy a Member-to-Leader Multi-Agent framework to extract and fuse features from multi-modal information.
We demonstrate the performance of our approach in both the 3D and 2D domains by employing the OS-MN40, OS-MN40-Miss, and Cifar10 datasets.
arXiv Detail & Related papers (2023-08-26T07:55:32Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - A Comparative Study of Data Augmentation Techniques for Deep Learning
Based Emotion Recognition [11.928873764689458]
We conduct a comprehensive evaluation of popular deep learning approaches for emotion recognition.
We show that long-range dependencies in the speech signal are critical for emotion recognition.
Speed/rate augmentation offers the most robust performance gain across models.
arXiv Detail & Related papers (2022-11-09T17:27:03Z) - EmailSum: Abstractive Email Thread Summarization [105.46012304024312]
We develop an abstractive Email Thread Summarization (EmailSum) dataset.
This dataset contains human-annotated short (30 words) and long (100 words) summaries of 2549 email threads.
Our results reveal the key challenges of current abstractive summarization models in this task.
arXiv Detail & Related papers (2021-07-30T15:13:14Z) - A Comprehensive Survey on Community Detection with Deep Learning [93.40332347374712]
A community reveals the features and connections of its members that are different from those in other communities in a network.
This survey devises and proposes a new taxonomy covering different categories of the state-of-the-art methods.
The main category, i.e., deep neural networks, is further divided into convolutional networks, graph attention networks, generative adversarial networks and autoencoders.
arXiv Detail & Related papers (2021-05-26T14:37:07Z) - Dataset Condensation with Gradient Matching [36.14340188365505]
We propose a training set synthesis technique for data-efficient learning, called dataset Condensation, that learns to condense large dataset into a small set of informative synthetic samples for training deep neural networks from scratch.
We rigorously evaluate its performance in several computer vision benchmarks and demonstrate that it significantly outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2020-06-10T16:30:52Z) - Distributed Learning in the Non-Convex World: From Batch to Streaming
Data, and Beyond [73.03743482037378]
Distributed learning has become a critical direction of the massively connected world envisioned by many.
This article discusses four key elements of scalable distributed processing and real-time data computation problems.
Practical issues and future research will also be discussed.
arXiv Detail & Related papers (2020-01-14T14:11:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.