Semantic Steganography: A Framework for Robust and High-Capacity   Information Hiding using Large Language Models
        - URL: http://arxiv.org/abs/2412.11043v1
- Date: Sun, 15 Dec 2024 04:04:23 GMT
- Title: Semantic Steganography: A Framework for Robust and High-Capacity   Information Hiding using Large Language Models
- Authors: Minhao Bai, Jinshuai Yang, Kaiyi Pang, Yongfeng Huang, Yue Gao, 
- Abstract summary: generative linguistic steganography has become a prevalent technique for hiding information within model-generated texts.<n>We propose a semantic steganography framework based on Large Language Models (LLMs)<n>This framework offers robustness and reliability for transmission in complex channels, as well as resistance to text rendering and word blocking.
- Score: 25.52890764952079
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract:   In the era of Large Language Models (LLMs), generative linguistic steganography has become a prevalent technique for hiding information within model-generated texts. However, traditional steganography methods struggle to effectively align steganographic texts with original model-generated texts due to the lower entropy of the predicted probability distribution of LLMs. This results in a decrease in embedding capacity and poses challenges for decoding stegos in real-world communication channels. To address these challenges, we propose a semantic steganography framework based on LLMs, which construct a semantic space and map secret messages onto this space using ontology-entity trees. This framework offers robustness and reliability for transmission in complex channels, as well as resistance to text rendering and word blocking. Additionally, the stegos generated by our framework are indistinguishable from the covers and achieve a higher embedding capacity compared to state-of-the-art steganography methods, while producing higher quality stegos. 
 
      
        Related papers
        - Robust Steganography from Large Language Models [1.5749416770494704]
 We study the problem of robust steganography.
We design and implement our steganographic schemes that embed arbitrary secret messages into natural language text.
 arXiv  Detail & Related papers  (2025-04-11T21:06:36Z)
- FreStega: A Plug-and-Play Method for Boosting Imperceptibility and   Capacity in Generative Linguistic Steganography for Real-World Scenarios [0.0]
 Linguistic steganography embeds secret information in seemingly innocent texts, safeguarding privacy in surveillance environments.
We propose FreStega, a plug-and-play method to reconstruct the distribution of language models used for generative linguistic steganography.
 arXiv  Detail & Related papers  (2024-12-27T13:56:51Z)
- Segment-Level Diffusion: A Framework for Controllable Long-Form   Generation with Diffusion Language Models [12.446047799880587]
 Token-level diffusion overlooks word-order dependencies and enforces short output windows, while passage-level diffusion struggles with learning robust representation for long-form text.<n>We propose Segment-Level Diffusion (SLD), a framework that enhances diffusion-based text generation through text segmentation, robust representation training with adversarial and contrastive learning, and improved latent-space guidance.
 arXiv  Detail & Related papers  (2024-12-15T22:47:44Z)
- HoliSDiP: Image Super-Resolution via Holistic Semantics and Diffusion   Prior [62.04939047885834]
 We present HoliSDiP, a framework that leverages semantic segmentation to provide both precise textual and spatial guidance for Real-ISR.
Our method employs semantic labels as concise text prompts while introducing dense semantic guidance through segmentation masks and our proposed spatial-CLIP Map.
 arXiv  Detail & Related papers  (2024-11-27T15:22:44Z)
- Decoding Diffusion: A Scalable Framework for Unsupervised Analysis of   Latent Space Biases and Representations Using Natural Language Prompts [68.48103545146127]
 This paper proposes a novel framework for unsupervised exploration of diffusion latent spaces.
We directly leverage natural language prompts and image captions to map latent directions.
Our method provides a more scalable and interpretable understanding of the semantic knowledge encoded within diffusion models.
 arXiv  Detail & Related papers  (2024-10-25T21:44:51Z)
- Unveiling Large Language Models Generated Texts: A Multi-Level   Fine-Grained Detection Framework [9.976099891796784]
 Large language models (LLMs) have transformed human writing by enhancing grammar correction, content expansion, and stylistic refinement.
Existing detection methods, which mainly rely on single-feature analysis and binary classification, often fail to effectively identify LLM-generated text in academic contexts.
We propose a novel Multi-level Fine-grained Detection framework that detects LLM-generated text by integrating low-level structural, high-level semantic, and deep-level linguistic features.
 arXiv  Detail & Related papers  (2024-10-18T07:25:00Z)
- MFCLIP: Multi-modal Fine-grained CLIP for Generalizable Diffusion Face   Forgery Detection [64.29452783056253]
 The rapid development of photo-realistic face generation methods has raised significant concerns in society and academia.
Although existing approaches mainly capture face forgery patterns using image modality, other modalities like fine-grained noises and texts are not fully explored.
We propose a novel multi-modal fine-grained CLIP (MFCLIP) model, which mines comprehensive and fine-grained forgery traces across image-noise modalities.
 arXiv  Detail & Related papers  (2024-09-15T13:08:59Z)
- ARTIST: Improving the Generation of Text-rich Images with Disentangled   Diffusion Models and Large Language Models [52.23899502520261]
 We introduce a novel framework named, ARTIST, which incorporates a dedicated textual diffusion model to focus on the learning of text structures specifically.<n>We finetune a visual diffusion model, enabling it to assimilate textual structure information from the pretrained textual model.<n>This disentangled architecture design and training strategy significantly enhance the text rendering ability of the diffusion models for text-rich image generation.
 arXiv  Detail & Related papers  (2024-06-17T19:31:24Z)
- Towards Next-Generation Steganalysis: LLMs Unleash the Power of   Detecting Steganography [18.7168443402118]
 Linguistic steganography provides convenient implementation to hide messages, particularly with the emergence of AI generation technology.
Existing methods are limited to finding distribution differences between steganographic texts and normal texts from the aspect of symbolic statistics.
This paper propose to employ human-like text processing abilities of large language models (LLMs) to realize the difference from the aspect of human perception.
 arXiv  Detail & Related papers  (2024-05-15T04:52:09Z)
- Learned representation-guided diffusion models for large-image   generation [58.192263311786824]
 We introduce a novel approach that trains diffusion models conditioned on embeddings from self-supervised learning (SSL)
Our diffusion models successfully project these features back to high-quality histopathology and remote sensing images.
Augmenting real data by generating variations of real images improves downstream accuracy for patch-level and larger, image-scale classification tasks.
 arXiv  Detail & Related papers  (2023-12-12T14:45:45Z)
- Disentangled Representation Learning with Large Language Models for
  Text-Attributed Graphs [57.052160123387104]
 We present the Disentangled Graph-Text Learner (DGTL) model, which is able to enhance the reasoning and predicting capabilities of LLMs for TAGs.
Our proposed DGTL model incorporates graph structure information through tailored disentangled graph neural network (GNN) layers.
 Experimental evaluations demonstrate the effectiveness of the proposed DGTL model on achieving superior or comparable performance over state-of-the-art baselines.
 arXiv  Detail & Related papers  (2023-10-27T14:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.