Advancing Email Spam Detection: Leveraging Zero-Shot Learning and Large Language Models
- URL: http://arxiv.org/abs/2505.02362v1
- Date: Mon, 05 May 2025 04:48:20 GMT
- Title: Advancing Email Spam Detection: Leveraging Zero-Shot Learning and Large Language Models
- Authors: Ghazaleh SHirvani, Saeid Ghasemshirazi,
- Abstract summary: This study investigates the effectiveness of Zero-Shot Learning using FLAN-T5 and advanced Natural Language Processing (NLP) techniques such as BERT for email spam detection.<n>The proposed approach aims to address the limitations of traditional spam detection systems.<n>The integration of FLAN-T5 and BERT enables robust spam detection without relying on extensive labeled datasets or frequent retraining.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Email spam detection is a critical task in modern communication systems, essential for maintaining productivity, security, and user experience. Traditional machine learning and deep learning approaches, while effective in static settings, face significant limitations in adapting to evolving spam tactics, addressing class imbalance, and managing data scarcity. These challenges necessitate innovative approaches that reduce dependency on extensive labeled datasets and frequent retraining. This study investigates the effectiveness of Zero-Shot Learning using FLAN-T5, combined with advanced Natural Language Processing (NLP) techniques such as BERT for email spam detection. By employing BERT to preprocess and extract critical information from email content, and FLAN-T5 to classify emails in a Zero-Shot framework, the proposed approach aims to address the limitations of traditional spam detection systems. The integration of FLAN-T5 and BERT enables robust spam detection without relying on extensive labeled datasets or frequent retraining, making it highly adaptable to unseen spam patterns and adversarial environments. This research highlights the potential of leveraging zero-shot learning and NLPs for scalable and efficient spam detection, providing insights into their capability to address the dynamic and challenging nature of spam detection tasks.
Related papers
- An Investigation of Large Language Models and Their Vulnerabilities in Spam Detection [7.550686419077825]
This project studies new spam detection systems that leverage Large Language Models (LLMs) fine-tuned with spam datasets.<n>This experimentation employs two LLM models of GPT2 and BERT and three spam datasets of Enron, LingSpam, and SMSspamCollection.<n>The results show that, while they can function as effective spam filters, the LLM models are susceptible to the adversarial and data poisoning attacks.
arXiv Detail & Related papers (2025-04-14T00:30:27Z) - SpaLLM-Guard: Pairing SMS Spam Detection Using Open-source and Commercial LLMs [1.3198171962008958]
We evaluate the potential of large language models (LLMs), both open-source and commercial, for SMS spam detection.<n>We compare their performance across zero-shot, few-shot, fine-tuning, and chain-of-thought prompting approaches.<n>Fine-tuning emerges as the most effective strategy, with Mixtral achieving 98.6% accuracy and a balanced false positive and false negative rate below 2%.
arXiv Detail & Related papers (2025-01-09T06:00:08Z) - Enhancing Web Spam Detection through a Blockchain-Enabled Crowdsourcing Mechanism [0.7303392100830282]
We propose blockchain-enabled incentivized crowdsourcing as a novel solution to enhance spam detection systems.
We create an incentive mechanism for data collection and labeling by leveraging blockchain's decentralized and transparent framework.
We show that incentivized crowdsourcing improves data quality, leading to more effective machine-learning models for spam detection.
arXiv Detail & Related papers (2024-10-01T16:53:42Z) - Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models [52.03511469562013]
We introduce the Iterative Contrastive Unlearning (ICU) framework, which consists of three core components.<n>A Knowledge Unlearning Induction module targets specific knowledge for removal using an unlearning loss.<n>A Contrastive Learning Enhancement module preserves the model's expressive capabilities against the pure unlearning goal.<n>An Iterative Unlearning Refinement module dynamically adjusts the unlearning process through ongoing evaluation and updates.
arXiv Detail & Related papers (2024-07-25T07:09:35Z) - Personalized Wireless Federated Learning for Large Language Models [75.22457544349668]
Large language models (LLMs) have driven profound transformations in wireless networks.<n>Within wireless environments, the training of LLMs faces significant challenges related to security and privacy.<n>This paper presents a systematic analysis of the training stages of LLMs in wireless networks, including pre-training, instruction tuning, and alignment tuning.
arXiv Detail & Related papers (2024-04-20T02:30:21Z) - Evaluating the Performance of ChatGPT for Spam Email Detection [15.690834818368929]
This study attempts to evaluate ChatGPT's capabilities for spam identification in both English and Chinese email datasets.<n>We employ ChatGPT for spam email detection using in-context learning, which requires a prompt instruction with (or without) a few demonstrations.<n>We also investigate how the number of demonstrations in the prompt affects the performance of ChatGPT.
arXiv Detail & Related papers (2024-02-23T04:52:08Z) - An Improved Transformer-based Model for Detecting Phishing, Spam, and
Ham: A Large Language Model Approach [0.0]
We present IPSDM, our model based on fine-tuning the BERT family of models to specifically detect phishing and spam email.
We demonstrate our fine-tuned version, IPSDM, is able to better classify emails in both unbalanced and balanced datasets.
arXiv Detail & Related papers (2023-11-01T18:41:50Z) - DST-Det: Simple Dynamic Self-Training for Open-Vocabulary Object Detection [72.25697820290502]
This work introduces a straightforward and efficient strategy to identify potential novel classes through zero-shot classification.
We refer to this approach as the self-training strategy, which enhances recall and accuracy for novel classes without requiring extra annotations, datasets, and re-training.
Empirical evaluations on three datasets, including LVIS, V3Det, and COCO, demonstrate significant improvements over the baseline performance.
arXiv Detail & Related papers (2023-10-02T17:52:24Z) - Deep convolutional forest: a dynamic deep ensemble approach for spam
detection in text [219.15486286590016]
This paper introduces a dynamic deep ensemble model for spam detection that adjusts its complexity and extracts features automatically.
As a result, the model achieved high precision, recall, f1-score and accuracy of 98.38%.
arXiv Detail & Related papers (2021-10-10T17:19:37Z) - Improving the Accuracy of Early Exits in Multi-Exit Architectures via
Curriculum Learning [88.17413955380262]
Multi-exit architectures allow deep neural networks to terminate their execution early in order to adhere to tight deadlines at the cost of accuracy.
We introduce a novel method called Multi-Exit Curriculum Learning that utilizes curriculum learning.
Our method consistently improves the accuracy of early exits compared to the standard training approach.
arXiv Detail & Related papers (2021-04-21T11:12:35Z) - Few-Cost Salient Object Detection with Adversarial-Paced Learning [95.0220555274653]
This paper proposes to learn the effective salient object detection model based on the manual annotation on a few training images only.
We name this task as the few-cost salient object detection and propose an adversarial-paced learning (APL)-based framework to facilitate the few-cost learning scenario.
arXiv Detail & Related papers (2021-04-05T14:15:49Z) - CosSGD: Nonlinear Quantization for Communication-efficient Federated
Learning [62.65937719264881]
Federated learning facilitates learning across clients without transferring local data on these clients to a central server.
We propose a nonlinear quantization for compressed gradient descent, which can be easily utilized in federated learning.
Our system significantly reduces the communication cost by up to three orders of magnitude, while maintaining convergence and accuracy of the training process.
arXiv Detail & Related papers (2020-12-15T12:20:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.