Cream Skimming the Underground: Identifying Relevant Information Points
from Online Forums
- URL: http://arxiv.org/abs/2308.02581v1
- Date: Thu, 3 Aug 2023 16:52:42 GMT
- Title: Cream Skimming the Underground: Identifying Relevant Information Points
from Online Forums
- Authors: Felipe Moreno-Vera, Mateus Nogueira, Cain\~a Figueiredo, Daniel Sadoc
Menasch\'e, Miguel Bicudo, Ashton Woiwood, Enrico Lovat, Anton Kocheturov,
Leandro Pfleger de Aguiar
- Abstract summary: This paper proposes a machine learning-based approach for detecting the exploitation of vulnerabilities in the wild by monitoring underground hacking forums.
We develop a supervised machine learning model that can filter threads citing CVEs and label them as Proof-of-Concept, Weaponization, or Exploitation.
- Score: 0.16252563723817934
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper proposes a machine learning-based approach for detecting the
exploitation of vulnerabilities in the wild by monitoring underground hacking
forums. The increasing volume of posts discussing exploitation in the wild
calls for an automatic approach to process threads and posts that will
eventually trigger alarms depending on their content. To illustrate the
proposed system, we use the CrimeBB dataset, which contains data scraped from
multiple underground forums, and develop a supervised machine learning model
that can filter threads citing CVEs and label them as Proof-of-Concept,
Weaponization, or Exploitation. Leveraging random forests, we indicate that
accuracy, precision and recall above 0.99 are attainable for the classification
task. Additionally, we provide insights into the difference in nature between
weaponization and exploitation, e.g., interpreting the output of a decision
tree, and analyze the profits and other aspects related to the hacking
communities. Overall, our work sheds insight into the exploitation of
vulnerabilities in the wild and can be used to provide additional ground truth
to models such as EPSS and Expected Exploitability.
Related papers
- "Glue pizza and eat rocks" -- Exploiting Vulnerabilities in Retrieval-Augmented Generative Models [74.05368440735468]
Retrieval-Augmented Generative (RAG) models enhance Large Language Models (LLMs)
In this paper, we demonstrate a security threat where adversaries can exploit the openness of these knowledge bases.
arXiv Detail & Related papers (2024-06-26T05:36:23Z) - Inferring Discussion Topics about Exploitation of Vulnerabilities from Underground Hacking Forums [0.0]
Underground hacking forums serve as breeding grounds for the exchange of hacking techniques and discussions related to exploitation.
We propose an innovative approach using topic modeling to analyze and uncover key themes in vulnerabilities discussed within these forums.
arXiv Detail & Related papers (2024-05-07T14:54:32Z) - Semi-supervised Open-World Object Detection [74.95267079505145]
We introduce a more realistic formulation, named semi-supervised open-world detection (SS-OWOD)
We demonstrate that the performance of the state-of-the-art OWOD detector dramatically deteriorates in the proposed SS-OWOD setting.
Our experiments on 4 datasets including MS COCO, PASCAL, Objects365 and DOTA demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-02-25T07:12:51Z) - Towards General Visual-Linguistic Face Forgery Detection [95.73987327101143]
Deepfakes are realistic face manipulations that can pose serious threats to security, privacy, and trust.
Existing methods mostly treat this task as binary classification, which uses digital labels or mask signals to train the detection model.
We propose a novel paradigm named Visual-Linguistic Face Forgery Detection(VLFFD), which uses fine-grained sentence-level prompts as the annotation.
arXiv Detail & Related papers (2023-07-31T10:22:33Z) - Free Lunch for Generating Effective Outlier Supervision [46.37464572099351]
We propose an ultra-effective method to generate near-realistic outlier supervision.
Our proposed textttBayesAug significantly reduces the false positive rate over 12.50% compared with the previous schemes.
arXiv Detail & Related papers (2023-01-17T01:46:45Z) - Explainable Abuse Detection as Intent Classification and Slot Filling [66.80201541759409]
We introduce the concept of policy-aware abuse detection, abandoning the unrealistic expectation that systems can reliably learn which phenomena constitute abuse from inspecting the data alone.
We show how architectures for intent classification and slot filling can be used for abuse detection, while providing a rationale for model decisions.
arXiv Detail & Related papers (2022-10-06T03:33:30Z) - CC-Cert: A Probabilistic Approach to Certify General Robustness of
Neural Networks [58.29502185344086]
In safety-critical machine learning applications, it is crucial to defend models against adversarial attacks.
It is important to provide provable guarantees for deep learning models against semantically meaningful input transformations.
We propose a new universal probabilistic certification approach based on Chernoff-Cramer bounds.
arXiv Detail & Related papers (2021-09-22T12:46:04Z) - Information Obfuscation of Graph Neural Networks [96.8421624921384]
We study the problem of protecting sensitive attributes by information obfuscation when learning with graph structured data.
We propose a framework to locally filter out pre-determined sensitive attributes via adversarial training with the total variation and the Wasserstein distance.
arXiv Detail & Related papers (2020-09-28T17:55:04Z) - Certifying Decision Trees Against Evasion Attacks by Program Analysis [9.290879387995401]
We propose a novel technique to verify the security of machine learning models against evasion attacks.
Our approach exploits the interpretability property of decision trees to transform them into imperative programs.
Our experiments show that our technique is both precise and efficient, yielding only a minimal number of false positives.
arXiv Detail & Related papers (2020-07-06T14:18:10Z) - Adaptive Double-Exploration Tradeoff for Outlier Detection [31.428683644520046]
We study a variant of the thresholding bandit problem (TBP) in the context of outlier detection.
The objective is to identify the outliers whose rewards are above a threshold.
By automatically trading off exploring the individual arms and exploring the outlier threshold, we provide an efficient algorithm.
arXiv Detail & Related papers (2020-05-13T00:12:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.