TIACam: Text-Anchored Invariant Feature Learning with Auto-Augmentation for Camera-Robust Zero-Watermarking
- URL: http://arxiv.org/abs/2602.18863v1
- Date: Sat, 21 Feb 2026 15:06:16 GMT
- Title: TIACam: Text-Anchored Invariant Feature Learning with Auto-Augmentation for Camera-Robust Zero-Watermarking
- Authors: Abdullah All Tanvir, Agnibh Dasgupta, Xin Zhong,
- Abstract summary: TIACam is a text-anchored invariant feature learning framework with auto-augmentation for camera-robust zero-watermarking.<n>Experiments on both synthetic and real-world camera captures demonstrate that TIACam achieves feature stability and watermark extraction accuracy.
- Score: 0.5429166905724048
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Camera recapture introduces complex optical degradations, such as perspective warping, illumination shifts, and Moiré interference, that remain challenging for deep watermarking systems. We present TIACam, a text-anchored invariant feature learning framework with auto-augmentation for camera-robust zero-watermarking. The method integrates three key innovations: (1) a learnable auto-augmentor that discovers camera-like distortions through differentiable geometric, photometric, and Moiré operators; (2) a text-anchored invariant feature learner that enforces semantic consistency via cross-modal adversarial alignment between image and text; and (3) a zero-watermarking head that binds binary messages in the invariant feature space without modifying image pixels. This unified formulation jointly optimizes invariance, semantic alignment, and watermark recoverability. Extensive experiments on both synthetic and real-world camera captures demonstrate that TIACam achieves state-of-the-art feature stability and watermark extraction accuracy, establishing a principled bridge between multimodal invariance learning and physically robust zero-watermarking.
Related papers
- RAVEN: Erasing Invisible Watermarks via Novel View Synthesis [35.417500510522835]
In this work, we expose a fundamental vulnerability in invisible watermarks by reformulating watermark removal as a view synthesis problem.<n>Our key insight is that generating a perceptually consistent alternative view of the same semantic content, naturally removes the embedded watermark while preserving visual fidelity.<n>We introduce a zero-shot diffusion-based framework that applies controlled geometric transformations in latent space, augmented with view-guided correspondence attention to maintain structural consistency during reconstruction.
arXiv Detail & Related papers (2026-01-13T18:59:58Z) - Pixel Seal: Adversarial-only training for invisible image and video watermarking [43.360750005378954]
Invisible watermarking is essential for tracing the provenance of digital content.<n>Current approaches often struggle to balance robustness against true imperceptibility.<n>This work introduces Pixel Seal, which sets a new state-of-the-art for image and video watermarking.
arXiv Detail & Related papers (2025-12-18T18:42:19Z) - InvZW: Invariant Feature Learning via Noise-Adversarial Training for Robust Image Zero-Watermarking [1.4042211166197214]
This paper introduces a novel deep learning framework for robust image zero-watermarking based on distortion-invariant feature learning.<n>As a zero-watermarking scheme, our method leaves the original image unaltered and learns a reference signature through optimization in the feature space.
arXiv Detail & Related papers (2025-06-25T12:32:08Z) - MorphMark: Flexible Adaptive Watermarking for Large Language Models [49.3302421751894]
Existing watermark methods often struggle with a dilemma: improving watermark effectiveness comes at the cost of reduced text quality.<n>We develop MorphMark method that adaptively adjusts the watermark strength in response to changes in the identified factor.<n>MorphMark achieves a superior resolution of the effectiveness-quality dilemma, while also offering greater flexibility and time and space efficiency.
arXiv Detail & Related papers (2025-05-14T13:11:16Z) - Text-Guided Image Invariant Feature Learning for Robust Image Watermarking [1.4042211166197214]
We propose a novel text-guided invariant feature learning framework for robust image watermarking.<n>We evaluate the proposed method across multiple datasets, demonstrating superior robustness against various image transformations.
arXiv Detail & Related papers (2025-03-18T01:32:38Z) - Fine-grained Image-to-LiDAR Contrastive Distillation with Visual Foundation Models [55.99654128127689]
Visual Foundation Models (VFMs) are used to generate semantic labels for weakly-supervised pixel-to-point contrastive distillation.<n>We adapt sampling probabilities of points to address imbalances in spatial distribution and category frequency.<n>Our approach consistently surpasses existing image-to-LiDAR contrastive distillation methods in downstream tasks.
arXiv Detail & Related papers (2024-05-23T07:48:19Z) - T2IW: Joint Text to Image & Watermark Generation [74.20148555503127]
We introduce a novel task for the joint generation of text to image and watermark (T2IW)
This T2IW scheme ensures minimal damage to image quality when generating a compound image by forcing the semantic feature and the watermark signal to be compatible in pixels.
We demonstrate remarkable achievements in image quality, watermark invisibility, and watermark robustness, supported by our proposed set of evaluation metrics.
arXiv Detail & Related papers (2023-09-07T16:12:06Z) - WMFormer++: Nested Transformer for Visible Watermark Removal via Implict
Joint Learning [68.00975867932331]
Existing watermark removal methods mainly rely on UNet with task-specific decoder branches.
We introduce an implicit joint learning paradigm to holistically integrate information from both branches.
The results demonstrate our approach's remarkable superiority, surpassing existing state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2023-08-20T07:56:34Z) - Exposure Fusion for Hand-held Camera Inputs with Optical Flow and
PatchMatch [53.149395644547226]
We propose a hybrid synthesis method for multi-exposure image fusion taken by hand-held cameras.
Our method can deal with such motions and maintain the exposure information of each input effectively.
Experiment results demonstrate the effectiveness and robustness of our method.
arXiv Detail & Related papers (2023-04-10T09:06:37Z) - Self-supervised Character-to-Character Distillation for Text Recognition [54.12490492265583]
We propose a novel self-supervised Character-to-Character Distillation method, CCD, which enables versatile augmentations to facilitate text representation learning.
CCD achieves state-of-the-art results, with average performance gains of 1.38% in text recognition, 1.7% in text segmentation, 0.24 dB (PSNR) and 0.0321 (SSIM) in text super-resolution.
arXiv Detail & Related papers (2022-11-01T05:48:18Z) - Degradation-agnostic Correspondence from Resolution-asymmetric Stereo [96.03964515969652]
We study the problem of stereo matching from a pair of images with different resolutions, e.g., those acquired with a tele-wide camera system.
We propose to impose the consistency between two views in a feature space instead of the image space, named feature-metric consistency.
We find that, although a stereo matching network trained with the photometric loss is not optimal, its feature extractor can produce degradation-agnostic and matching-specific features.
arXiv Detail & Related papers (2022-04-04T12:24:34Z) - GenRadar: Self-supervised Probabilistic Camera Synthesis based on Radar
Frequencies [12.707035083920227]
This work combines the complementary strengths of both sensor types in a unique self-learning fusion approach for a probabilistic scene reconstruction.
A proposed algorithm exploits similarities and establishes correspondences between both domains at different feature levels during training.
These discrete tokens are finally transformed back into an instructive view of the respective surrounding, allowing to visually perceive potential dangers.
arXiv Detail & Related papers (2021-07-19T15:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.