DeepTaster: Adversarial Perturbation-Based Fingerprinting to Identify
Proprietary Dataset Use in Deep Neural Networks
- URL: http://arxiv.org/abs/2211.13535v2
- Date: Thu, 4 Jan 2024 00:16:22 GMT
- Title: DeepTaster: Adversarial Perturbation-Based Fingerprinting to Identify
Proprietary Dataset Use in Deep Neural Networks
- Authors: Seonhye Park, Alsharif Abuadbba, Shuo Wang, Kristen Moore, Yansong
Gao, Hyoungshick Kim, Surya Nepal
- Abstract summary: We introduce DeepTaster, a novel fingerprinting technique to address scenarios where a victim's data is unlawfully used to build a suspect model.
To accomplish this, DeepTaster generates adversarial images with perturbations, transforms them into the Fourier frequency domain, and uses these transformed images to identify the dataset used in a suspect model.
- Score: 34.11970637801044
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Training deep neural networks (DNNs) requires large datasets and powerful
computing resources, which has led some owners to restrict redistribution
without permission. Watermarking techniques that embed confidential data into
DNNs have been used to protect ownership, but these can degrade model
performance and are vulnerable to watermark removal attacks. Recently,
DeepJudge was introduced as an alternative approach to measuring the similarity
between a suspect and a victim model. While DeepJudge shows promise in
addressing the shortcomings of watermarking, it primarily addresses situations
where the suspect model copies the victim's architecture. In this study, we
introduce DeepTaster, a novel DNN fingerprinting technique, to address
scenarios where a victim's data is unlawfully used to build a suspect model.
DeepTaster can effectively identify such DNN model theft attacks, even when the
suspect model's architecture deviates from the victim's. To accomplish this,
DeepTaster generates adversarial images with perturbations, transforms them
into the Fourier frequency domain, and uses these transformed images to
identify the dataset used in a suspect model. The underlying premise is that
adversarial images can capture the unique characteristics of DNNs built with a
specific dataset. To demonstrate the effectiveness of DeepTaster, we evaluated
the effectiveness of DeepTaster by assessing its detection accuracy on three
datasets (CIFAR10, MNIST, and Tiny-ImageNet) across three model architectures
(ResNet18, VGG16, and DenseNet161). We conducted experiments under various
attack scenarios, including transfer learning, pruning, fine-tuning, and data
augmentation. Specifically, in the Multi-Architecture Attack scenario,
DeepTaster was able to identify all the stolen cases across all datasets, while
DeepJudge failed to detect any of the cases.
Related papers
- DeepiSign-G: Generic Watermark to Stamp Hidden DNN Parameters for Self-contained Tracking [15.394110881491773]
DeepiSign-G is a versatile watermarking approach designed for comprehensive verification of leading DNN architectures, including CNNs and RNNs.
Unlike traditional hashing techniques, DeepiSign-G allows substantial metadata incorporation directly within the model, enabling detailed, self-contained tracking and verification.
We demonstrate DeepiSign-G's applicability across various architectures, including CNN models (VGG, ResNets, DenseNet) and RNNs (Text sentiment classifiers)
arXiv Detail & Related papers (2024-07-01T13:15:38Z) - TEN-GUARD: Tensor Decomposition for Backdoor Attack Detection in Deep
Neural Networks [3.489779105594534]
We introduce a novel approach to backdoor detection using two tensor decomposition methods applied to network activations.
This has a number of advantages relative to existing detection methods, including the ability to analyze multiple models at the same time.
Results show that our method detects backdoored networks more accurately and efficiently than current state-of-the-art methods.
arXiv Detail & Related papers (2024-01-06T03:08:28Z) - A Geometrical Approach to Evaluate the Adversarial Robustness of Deep
Neural Networks [52.09243852066406]
Adversarial Converging Time Score (ACTS) measures the converging time as an adversarial robustness metric.
We validate the effectiveness and generalization of the proposed ACTS metric against different adversarial attacks on the large-scale ImageNet dataset.
arXiv Detail & Related papers (2023-10-10T09:39:38Z) - Safe and Robust Watermark Injection with a Single OoD Image [90.71804273115585]
Training a high-performance deep neural network requires large amounts of data and computational resources.
We propose a safe and robust backdoor-based watermark injection technique.
We induce random perturbation of model parameters during watermark injection to defend against common watermark removal attacks.
arXiv Detail & Related papers (2023-09-04T19:58:35Z) - On the Adversarial Inversion of Deep Biometric Representations [3.804240190982696]
Biometric authentication service providers often claim that it is not possible to reverse-engineer a user's raw biometric sample.
In this paper, we investigate this claim on the specific example of deep neural network (DNN) embeddings.
We propose a two-pronged attack that first infers the original DNN by exploiting the model footprint on the embedding.
arXiv Detail & Related papers (2023-04-12T01:47:11Z) - Black-box Dataset Ownership Verification via Backdoor Watermarking [67.69308278379957]
We formulate the protection of released datasets as verifying whether they are adopted for training a (suspicious) third-party model.
We propose to embed external patterns via backdoor watermarking for the ownership verification to protect them.
Specifically, we exploit poison-only backdoor attacks ($e.g.$, BadNets) for dataset watermarking and design a hypothesis-test-guided method for dataset verification.
arXiv Detail & Related papers (2022-08-04T05:32:20Z) - DeepSteal: Advanced Model Extractions Leveraging Efficient Weight
Stealing in Memories [26.067920958354]
One of the major threats to the privacy of Deep Neural Networks (DNNs) is model extraction attacks.
Recent studies show hardware-based side channel attacks can reveal internal knowledge about DNN models (e.g., model architectures)
We propose an advanced model extraction attack framework DeepSteal that effectively steals DNN weights with the aid of memory side-channel attack.
arXiv Detail & Related papers (2021-11-08T16:55:45Z) - Exploring Structure Consistency for Deep Model Watermarking [122.38456787761497]
The intellectual property (IP) of Deep neural networks (DNNs) can be easily stolen'' by surrogate model attack.
We propose a new watermarking methodology, namely structure consistency'', based on which a new deep structure-aligned model watermarking algorithm is designed.
arXiv Detail & Related papers (2021-08-05T04:27:15Z) - Reversible Watermarking in Deep Convolutional Neural Networks for
Integrity Authentication [78.165255859254]
We propose a reversible watermarking algorithm for integrity authentication.
The influence of embedding reversible watermarking on the classification performance is less than 0.5%.
At the same time, the integrity of the model can be verified by applying the reversible watermarking.
arXiv Detail & Related papers (2021-04-09T09:32:21Z) - DeepiSign: Invisible Fragile Watermark to Protect the Integrityand
Authenticity of CNN [37.98139322456872]
We propose a self-contained tamper-proofing method, called DeepiSign, to ensure the integrity and authenticity of CNN models.
DeepiSign applies the idea of fragile invisible watermarking to securely embed a secret and its hash value into a CNN model.
Our theoretical analysis shows that DeepiSign can hide up to 1KB secret in each layer with minimal loss of the model's accuracy.
arXiv Detail & Related papers (2021-01-12T06:42:45Z) - Cooling-Shrinking Attack: Blinding the Tracker with Imperceptible Noises [87.53808756910452]
A cooling-shrinking attack method is proposed to deceive state-of-the-art SiameseRPN-based trackers.
Our method has good transferability and is able to deceive other top-performance trackers such as DaSiamRPN, DaSiamRPN-UpdateNet, and DiMP.
arXiv Detail & Related papers (2020-03-21T07:13:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.