StegaPos: Preventing Crops and Splices with Imperceptible Positional
Encodings
- URL: http://arxiv.org/abs/2104.12290v1
- Date: Sun, 25 Apr 2021 23:42:29 GMT
- Title: StegaPos: Preventing Crops and Splices with Imperceptible Positional
Encodings
- Authors: Gokhan Egri, Todd Zickler
- Abstract summary: We present a model for distinguishing between images that are authentic copies of ones published by photographers.
The model comprises an encoder that resides with the photographer and a matching decoder that is available to observers.
We find that training the encoder and decoder together produces a model that imperceptibly encodes position.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a model for differentiating between images that are authentic
copies of ones published by photographers, and images that have been
manipulated by cropping, splicing or downsampling after publication. The model
comprises an encoder that resides with the photographer and a matching decoder
that is available to observers. The encoder learns to embed imperceptible
positional signatures into image values prior to publication. The decoder
learns to use these steganographic positional ("stegapos") signatures to
determine, for each small image patch, the 2D positional coordinates that were
held by the patch in its originally-published image. Crop, splice and
downsample edits become detectable by the inconsistencies they cause in the
hidden positional signatures. We find that training the encoder and decoder
together produces a model that imperceptibly encodes position, and that enables
superior performance on established benchmarks for splice detection and high
accuracy on a new benchmark for crop detection.
Related papers
- LCM-Lookahead for Encoder-based Text-to-Image Personalization [82.56471486184252]
We explore the potential of using shortcut-mechanisms to guide the personalization of text-to-image models.
We focus on encoder-based personalization approaches, and demonstrate that by tuning them with a lookahead identity loss, we can achieve higher identity fidelity.
arXiv Detail & Related papers (2024-04-04T17:43:06Z) - StegGuard: Fingerprinting Self-supervised Pre-trained Encoders via Secrets Embeder and Extractor [20.88708866426335]
We propose StegGuard, a novel fingerprinting mechanism to verify the ownership of the suspect pre-trained encoder using steganography.
A critical perspective in StegGuard is that the unique characteristic of the transformation from an image to an embedding, conducted by the pre-trained encoder, can be equivalently exposed.
StegGuard can reliably identify across varied independent encoders, and is robust against model stealing related attacks.
arXiv Detail & Related papers (2023-10-05T08:30:42Z) - Neural Distributed Image Compression with Cross-Attention Feature
Alignment [1.2234742322758418]
We consider a pair of stereo images, which have overlapping fields of view, captured by a synchronized and calibrated pair of cameras.
We assume that one image of the pair is to be compressed and transmitted, while the other image is available only at the decoder.
In the proposed architecture, the encoder maps the input image to a latent space using a DNN, quantizes the latent representation, and compresses it losslessly using entropy coding.
arXiv Detail & Related papers (2022-07-18T10:15:04Z) - Context Autoencoder for Self-Supervised Representation Learning [64.63908944426224]
We pretrain an encoder by making predictions in the encoded representation space.
The network is an encoder-regressor-decoder architecture.
We demonstrate the effectiveness of our CAE through superior transfer performance in downstream tasks.
arXiv Detail & Related papers (2022-02-07T09:33:45Z) - DocEnTr: An End-to-End Document Image Enhancement Transformer [13.108797370734893]
Document images can be affected by many degradation scenarios, which cause recognition and processing difficulties.
We present a new encoder-decoder architecture based on vision transformers to enhance both machine-printed and handwritten document images.
arXiv Detail & Related papers (2022-01-25T11:45:35Z) - Compact Binary Fingerprint for Image Copy Re-Ranking [0.0]
Image copy detection is challenging and appealing topic in computer vision and signal processing.
Local keypoint descriptors such as SIFT are used to represent the images, and based on those descriptors matching, images are matched and retrieved.
Features are quantized so that searching/matching may be made feasible for large databases at the cost of accuracy loss.
arXiv Detail & Related papers (2021-09-16T08:44:56Z) - EncoderMI: Membership Inference against Pre-trained Encoders in
Contrastive Learning [27.54202989524394]
We proposeMI, the first membership inference method against image encoders pre-trained by contrastive learning.
We evaluateMI on image encoders pre-trained on multiple datasets by ourselves as well as the Contrastive Language-Image Pre-training (CLIP) image encoder, which is pre-trained on 400 million (image, text) pairs collected from the Internet and released by OpenAI.
arXiv Detail & Related papers (2021-08-25T03:00:45Z) - Learnable Fourier Features for Multi-DimensionalSpatial Positional
Encoding [96.9752763607738]
We propose a novel positional encoding method based on learnable Fourier features.
Our experiments show that our learnable feature representation for multi-dimensional positional encoding outperforms existing methods.
arXiv Detail & Related papers (2021-06-05T04:40:18Z) - D-Unet: A Dual-encoder U-Net for Image Splicing Forgery Detection and
Localization [108.8592577019391]
Image splicing forgery detection is a global binary classification task that distinguishes the tampered and non-tampered regions by image fingerprints.
We propose a novel network called dual-encoder U-Net (D-Unet) for image splicing forgery detection, which employs an unfixed encoder and a fixed encoder.
In an experimental comparison study of D-Unet and state-of-the-art methods, D-Unet outperformed the other methods in image-level and pixel-level detection.
arXiv Detail & Related papers (2020-12-03T10:54:02Z) - Swapping Autoencoder for Deep Image Manipulation [94.33114146172606]
We propose the Swapping Autoencoder, a deep model designed specifically for image manipulation.
The key idea is to encode an image with two independent components and enforce that any swapped combination maps to a realistic image.
Experiments on multiple datasets show that our model produces better results and is substantially more efficient compared to recent generative models.
arXiv Detail & Related papers (2020-07-01T17:59:57Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.