A Dynamic YOLO-Based Sequence-Matching Model for Efficient Coverless Image Steganography
- URL: http://arxiv.org/abs/2401.11946v1
- Date: Mon, 22 Jan 2024 13:35:27 GMT
- Title: A Dynamic YOLO-Based Sequence-Matching Model for Efficient Coverless Image Steganography
- Authors: Jiajun Liu, Lina Tan, Zhili Zhou, Yi Li, Peng Chen,
- Abstract summary: We present an efficient coverless steganography scheme based on dynamically matched images.
YOLO is employed for selecting optimal objects, and a mapping dictionary is established between these objects and scrambling factors.
Under typical geometric attacks, it can recover 79.85% of secret information on average.
- Score: 12.950354250175982
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many existing coverless steganography methods establish a mapping relationship between cover images and hidden data. There exists an issue that the number of images stored in the database grows exponentially as the steganographic capacity rises. The need for a high steganographic capacity makes it challenging to build an image database. To improve the image library utilization and anti-attack capability of the steganography system, we present an efficient coverless scheme based on dynamically matched substrings. YOLO is employed for selecting optimal objects, and a mapping dictionary is established between these objects and scrambling factors. With the aid of this dictionary, each image is effectively assigned to a specific scrambling factor, which is used to scramble the receiver's sequence key. To achieve sufficient steganography capability based on a limited image library, all substrings of the scrambled sequences hold the potential to hide data. After completing the secret information matching, the ideal number of stego images will be obtained from the database. According to experimental results, this technology outperforms most previous works on data load, transmission security, and hiding capacity. Under typical geometric attacks, it can recover 79.85\% of secret information on average. Furthermore, only approximately 200 random images are needed to meet a capacity of 19 bits per image.
Related papers
- StegaINR4MIH: steganography by implicit neural representation for multi-image hiding [6.29495604869364]
Multi-image hiding, which embeds multiple secret images into a cover image, has gradually become a research hotspot in the field of image steganography.
We propose StegaINR4MIH, a novel implicit neural representation steganography framework that enables the hiding of multiple images within a single implicit representation function.
arXiv Detail & Related papers (2024-10-14T03:09:41Z) - EmbAu: A Novel Technique to Embed Audio Data Using Shuffled Frog Leaping
Algorithm [0.7673339435080445]
The aim of steganographic algorithms is to identify the appropriate pixel positions in the host or cover image, where bits of sensitive information can be concealed for data encryption.
Work is being done to improve the capacity to integrate sensitive information and to maintain the visual appearance of the steganographic image.
We use the Shuffled Frog Leaping Algorithm (SFLA) to determine the order of pixels by which sensitive information can be placed in the cover image.
arXiv Detail & Related papers (2023-12-13T17:34:08Z) - Semantic Generative Augmentations for Few-Shot Counting [0.0]
We investigate how synthetic data can benefit few-shot class-agnostic counting.
We propose to rely on a double conditioning of Stable Diffusion with both a prompt and a density map.
Our experiments show that our diversified generation strategy significantly improves the counting accuracy of two recent and performing few-shot counting models.
arXiv Detail & Related papers (2023-10-26T11:42:48Z) - Improving Multimodal Datasets with Image Captioning [65.74736570293622]
We study how generated captions can increase the utility of web-scraped datapoints with nondescript text.
Our experiments with using generated captions at DataComp's large scale (1.28B image-text pairs) offer insights into the limitations of synthetic text.
arXiv Detail & Related papers (2023-07-19T17:47:12Z) - ConfounderGAN: Protecting Image Data Privacy with Causal Confounder [85.6757153033139]
We propose ConfounderGAN, a generative adversarial network (GAN) that can make personal image data unlearnable to protect the data privacy of its owners.
Experiments are conducted in six image classification datasets, consisting of three natural object datasets and three medical datasets.
arXiv Detail & Related papers (2022-12-04T08:49:14Z) - Weakly Supervised Annotations for Multi-modal Greeting Cards Dataset [8.397847537464534]
We propose to aggregate features from pretrained images and text embeddings to learn abstract visual concepts from Greeting Cards dataset.
The proposed dataset is also useful for generating greeting card images using pre-trained text-to-image generation model.
arXiv Detail & Related papers (2022-12-01T20:07:52Z) - Hiding Images in Deep Probabilistic Models [58.23127414572098]
We describe a different computational framework to hide images in deep probabilistic models.
Specifically, we use a DNN to model the probability density of cover images, and hide a secret image in one particular location of the learned distribution.
We demonstrate the feasibility of our SinGAN approach in terms of extraction accuracy and model security.
arXiv Detail & Related papers (2022-10-05T13:33:25Z) - Generative Steganography Network [37.182458848616754]
We propose an advanced generative steganography network (GSN) that can generate realistic stego images without using cover images.
A module named secret block is designed delicately to conceal secret data in the feature maps during image generation.
arXiv Detail & Related papers (2022-07-28T03:34:37Z) - Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone [170.85076677740292]
We present FIBER (Fusion-In-the-Backbone-basedER), a new model architecture for vision-language (VL) pre-training.
Instead of having dedicated transformer layers for fusion after the uni-modal backbones, FIBER pushes multimodal fusion deep into the model.
We conduct comprehensive experiments on a wide range of VL tasks, ranging from VQA, image captioning, and retrieval, to phrase grounding, referring expression comprehension, and object detection.
arXiv Detail & Related papers (2022-06-15T16:41:29Z) - Prefix Conditioning Unifies Language and Label Supervision [84.11127588805138]
We show that dataset biases negatively affect pre-training by reducing the generalizability of learned representations.
In experiments, we show that this simple technique improves the performance in zero-shot image recognition accuracy and robustness to the image-level distribution shift.
arXiv Detail & Related papers (2022-06-02T16:12:26Z) - Using Text to Teach Image Retrieval [47.72498265721957]
We build on the concept of image manifold to represent the feature space of images, learned via neural networks, as a graph.
We augment the manifold samples with geometrically aligned text, thereby using a plethora of sentences to teach us about images.
The experimental results show that the joint embedding manifold is a robust representation, allowing it to be a better basis to perform image retrieval.
arXiv Detail & Related papers (2020-11-19T16:09:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.