Steganographic Capacity of Deep Learning Models
- URL: http://arxiv.org/abs/2306.17189v1
- Date: Sun, 25 Jun 2023 13:43:35 GMT
- Title: Steganographic Capacity of Deep Learning Models
- Authors: Lei Zhang and Dong Li and Olha Jure\v{c}kov\'a and Mark Stamp
- Abstract summary: We consider the steganographic capacity of several learning models.
We train a Multilayer Perceptron (MLP), Convolutional Neural Network (CNN), and Transformer model on a challenging malware classification problem.
We find that the steganographic capacity of the learning models tested is surprisingly high, and that in each case, there is a clear threshold after which model performance rapidly degrades.
- Score: 12.974139332068491
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As machine learning and deep learning models become ubiquitous, it is
inevitable that there will be attempts to exploit such models in various attack
scenarios. For example, in a steganographic-based attack, information could be
hidden in a learning model, which might then be used to distribute malware, or
for other malicious purposes. In this research, we consider the steganographic
capacity of several learning models. Specifically, we train a Multilayer
Perceptron (MLP), Convolutional Neural Network (CNN), and Transformer model on
a challenging malware classification problem. For each of the resulting models,
we determine the number of low-order bits of the trained parameters that can be
altered without significantly affecting the performance of the model. We find
that the steganographic capacity of the learning models tested is surprisingly
high, and that in each case, there is a clear threshold after which model
performance rapidly degrades.
Related papers
- Reinforcing Pre-trained Models Using Counterfactual Images [54.26310919385808]
This paper proposes a novel framework to reinforce classification models using language-guided generated counterfactual images.
We identify model weaknesses by testing the model using the counterfactual image dataset.
We employ the counterfactual images as an augmented dataset to fine-tune and reinforce the classification model.
arXiv Detail & Related papers (2024-06-19T08:07:14Z) - Identifying and Mitigating Model Failures through Few-shot CLIP-aided
Diffusion Generation [65.268245109828]
We propose an end-to-end framework to generate text descriptions of failure modes associated with spurious correlations.
These descriptions can be used to generate synthetic data using generative models, such as diffusion models.
Our experiments have shown remarkable textbfimprovements in accuracy ($sim textbf21%$) on hard sub-populations.
arXiv Detail & Related papers (2023-12-09T04:43:49Z) - On the Steganographic Capacity of Selected Learning Models [1.0640226829362012]
We consider the question of the steganographic capacity of learning models.
For a wide range of models, we determine the number of low-order bits that can be overwritten.
Of the models tested, the steganographic capacity ranges from 7.04 KB for our LR experiments, to 44.74 MB for InceptionV3.
arXiv Detail & Related papers (2023-08-29T10:41:34Z) - Learning to Jump: Thinning and Thickening Latent Counts for Generative
Modeling [69.60713300418467]
Learning to jump is a general recipe for generative modeling of various types of data.
We demonstrate when learning to jump is expected to perform comparably to learning to denoise, and when it is expected to perform better.
arXiv Detail & Related papers (2023-05-28T05:38:28Z) - An Empirical Study of Deep Learning Models for Vulnerability Detection [4.243592852049963]
We surveyed and reproduced 9 state-of-the-art deep learning models on 2 widely used vulnerability detection datasets.
We investigated model capabilities, training data, and model interpretation.
Our findings can help better understand model results, provide guidance on preparing training data, and improve the robustness of the models.
arXiv Detail & Related papers (2022-12-15T19:49:34Z) - A Distillation Learning Model of Adaptive Structural Deep Belief Network
for AffectNet: Facial Expression Image Database [0.0]
We have developed the adaptive structure learning method of Deep Belief Network (DBN)
In this paper, our model is applied to a facial expression image data set, AffectNet.
The classification accuracy was improved from 78.4% to 91.3% by the proposed method.
arXiv Detail & Related papers (2021-10-25T08:01:36Z) - ViViT: A Video Vision Transformer [75.74690759089529]
We present pure-transformer based models for video classification.
Our model extracts-temporal tokens from the input video, which are then encoded by a series of transformer layers.
We show how we can effectively regularise the model during training and leverage pretrained image models to be able to train on comparatively small datasets.
arXiv Detail & Related papers (2021-03-29T15:27:17Z) - Firearm Detection via Convolutional Neural Networks: Comparing a
Semantic Segmentation Model Against End-to-End Solutions [68.8204255655161]
Threat detection of weapons and aggressive behavior from live video can be used for rapid detection and prevention of potentially deadly incidents.
One way for achieving this is through the use of artificial intelligence and, in particular, machine learning for image analysis.
We compare a traditional monolithic end-to-end deep learning model and a previously proposed model based on an ensemble of simpler neural networks detecting fire-weapons via semantic segmentation.
arXiv Detail & Related papers (2020-12-17T15:19:29Z) - On the Transferability of Adversarial Attacksagainst Neural Text
Classifier [121.6758865857686]
We investigate the transferability of adversarial examples for text classification models.
We propose a genetic algorithm to find an ensemble of models that can induce adversarial examples to fool almost all existing models.
We derive word replacement rules that can be used for model diagnostics from these adversarial examples.
arXiv Detail & Related papers (2020-11-17T10:45:05Z) - Editable Neural Networks [25.939872732737022]
In many applications, a single model error can lead to devastating financial, reputational and even life-threatening consequences.
We propose Editable Training, a model-agnostic training technique that encourages fast editing of the trained model.
We empirically demonstrate the effectiveness of this method on large-scale image classification and machine translation tasks.
arXiv Detail & Related papers (2020-04-01T11:26:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.