SSL-WM: A Black-Box Watermarking Approach for Encoders Pre-trained by
Self-supervised Learning
- URL: http://arxiv.org/abs/2209.03563v2
- Date: Mon, 29 Jan 2024 14:25:13 GMT
- Title: SSL-WM: A Black-Box Watermarking Approach for Encoders Pre-trained by
Self-supervised Learning
- Authors: Peizhuo Lv, Pan Li, Shenchen Zhu, Shengzhi Zhang, Kai Chen, Ruigang
Liang, Chang Yue, Fan Xiang, Yuling Cai, Hualong Ma, Yingjun Zhang, Guozhu
Meng
- Abstract summary: Self-Supervised Learning (SSL) has been widely utilized to facilitate various downstream tasks in Computer Vision (CV) and Natural Language Processing (NLP) domains.
attackers may steal such SSL models and commercialize them for profit, making it crucial to verify the ownership of the SSL models.
Most existing ownership protection solutions (e.g., backdoor-based watermarks) are designed for supervised learning models.
We propose a novel black-box watermarking solution, named SSL-WM, for verifying the ownership of SSL models.
- Score: 21.36798280084255
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent years have witnessed tremendous success in Self-Supervised Learning
(SSL), which has been widely utilized to facilitate various downstream tasks in
Computer Vision (CV) and Natural Language Processing (NLP) domains. However,
attackers may steal such SSL models and commercialize them for profit, making
it crucial to verify the ownership of the SSL models. Most existing ownership
protection solutions (e.g., backdoor-based watermarks) are designed for
supervised learning models and cannot be used directly since they require that
the models' downstream tasks and target labels be known and available during
watermark embedding, which is not always possible in the domain of SSL. To
address such a problem, especially when downstream tasks are diverse and
unknown during watermark embedding, we propose a novel black-box watermarking
solution, named SSL-WM, for verifying the ownership of SSL models. SSL-WM maps
watermarked inputs of the protected encoders into an invariant representation
space, which causes any downstream classifier to produce expected behavior,
thus allowing the detection of embedded watermarks. We evaluate SSL-WM on
numerous tasks, such as CV and NLP, using different SSL models both
contrastive-based and generative-based. Experimental results demonstrate that
SSL-WM can effectively verify the ownership of stolen SSL models in various
downstream tasks. Furthermore, SSL-WM is robust against model fine-tuning,
pruning, and input preprocessing attacks. Lastly, SSL-WM can also evade
detection from evaluated watermark detection approaches, demonstrating its
promising application in protecting the ownership of SSL models.
Related papers
- Erasing the Bias: Fine-Tuning Foundation Models for Semi-Supervised Learning [4.137391543972184]
Semi-supervised learning (SSL) has witnessed remarkable progress, resulting in numerous method variations.
In this paper, we present a novel SSL approach named FineSSL that significantly addresses this limitation by adapting pre-trained foundation models.
We demonstrate that FineSSL sets a new state of the art for SSL on multiple benchmark datasets, reduces the training cost by over six times, and can seamlessly integrate various fine-tuning and modern SSL algorithms.
arXiv Detail & Related papers (2024-05-20T03:33:12Z) - Towards Adversarial Robustness And Backdoor Mitigation in SSL [0.562479170374811]
Self-Supervised Learning (SSL) has shown great promise in learning representations from unlabeled data.
SSL methods have recently been shown to be vulnerable to backdoor attacks.
This work aims to address defending against backdoor attacks in SSL.
arXiv Detail & Related papers (2024-03-23T19:21:31Z) - Erasing Self-Supervised Learning Backdoor by Cluster Activation Masking [65.44477004525231]
Researchers have recently found that Self-Supervised Learning (SSL) is vulnerable to backdoor attacks.
In this paper, we propose to erase the SSL backdoor by cluster activation masking and propose a novel PoisonCAM method.
Our method achieves 96% accuracy for backdoor trigger detection compared to 3% of the state-of-the-art method on poisoned ImageNet-100.
arXiv Detail & Related papers (2023-12-13T08:01:15Z) - Self-Supervision for Tackling Unsupervised Anomaly Detection: Pitfalls
and Opportunities [50.231837687221685]
Self-supervised learning (SSL) has transformed machine learning and its many real world applications.
Unsupervised anomaly detection (AD) has also capitalized on SSL, by self-generating pseudo-anomalies.
arXiv Detail & Related papers (2023-08-28T07:55:01Z) - SSL-Cleanse: Trojan Detection and Mitigation in Self-Supervised Learning [27.68997463681079]
Self-supervised learning (SSL) is a prevalent approach for encoding data representations.
Trojan attacks embedded in SSL encoders can operate covertly, spreading across multiple users and devices.
We introduce SSL-Cleanse as a solution to identify and mitigate backdoor threats in SSL encoders.
arXiv Detail & Related papers (2023-03-16T04:45:06Z) - A Survey on Self-supervised Learning: Algorithms, Applications, and Future Trends [82.64268080902742]
Self-supervised learning (SSL) aims to learn discriminative features from unlabeled data without relying on human-annotated labels.
SSL has garnered significant attention recently, leading to the development of numerous related algorithms.
This paper presents a review of diverse SSL methods, encompassing algorithmic aspects, application domains, three key trends, and open research questions.
arXiv Detail & Related papers (2023-01-13T14:41:05Z) - A Survey on Masked Autoencoder for Self-supervised Learning in Vision
and Beyond [64.85076239939336]
Self-supervised learning (SSL) in vision might undertake a similar trajectory as in NLP.
generative pretext tasks with the masked prediction (e.g., BERT) have become a de facto standard SSL practice in NLP.
Success of mask image modeling has revived the masking autoencoder.
arXiv Detail & Related papers (2022-07-30T09:59:28Z) - Combining Spectral and Self-Supervised Features for Low Resource Speech
Recognition and Translation [27.857955394020475]
Self-Supervised Learning (SSL) models have been successfully applied in various deep learning-based speech tasks.
The quality of SSL representations depends highly on the relatedness between the SSL training domain(s) and the target data domain.
We propose a learnable and interpretable framework to combine SF and SSL representations.
arXiv Detail & Related papers (2022-04-05T20:09:15Z) - DATA: Domain-Aware and Task-Aware Pre-training [94.62676913928831]
We present DATA, a simple yet effective NAS approach specialized for self-supervised learning (SSL)
Our method achieves promising results across a wide range of computation costs on downstream tasks, including image classification, object detection and semantic segmentation.
arXiv Detail & Related papers (2022-03-17T02:38:49Z) - Sound and Visual Representation Learning with Multiple Pretraining Tasks [104.11800812671953]
Self-supervised tasks (SSL) reveal different features from the data.
This work aims to combine Multiple SSL tasks (Multi-SSL) that generalizes well for all downstream tasks.
Experiments on sound representations demonstrate that Multi-SSL via incremental learning (IL) of SSL tasks outperforms single SSL task models.
arXiv Detail & Related papers (2022-01-04T09:09:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.