FaceLeaks: Inference Attacks against Transfer Learning Models via
Black-box Queries
- URL: http://arxiv.org/abs/2010.14023v1
- Date: Tue, 27 Oct 2020 03:02:40 GMT
- Title: FaceLeaks: Inference Attacks against Transfer Learning Models via
Black-box Queries
- Authors: Seng Pei Liew and Tsubasa Takahashi
- Abstract summary: We investigate if one can leak or infer private information without interacting with the teacher model directly.
We propose novel strategies to infer from aggregate-level information.
Our study indicates that information leakage is a real privacy threat to the transfer learning framework widely used in real-life situations.
- Score: 2.7564955518050693
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transfer learning is a useful machine learning framework that allows one to
build task-specific models (student models) without significantly incurring
training costs using a single powerful model (teacher model) pre-trained with a
large amount of data. The teacher model may contain private data, or interact
with private inputs. We investigate if one can leak or infer such private
information without interacting with the teacher model directly. We describe
such inference attacks in the context of face recognition, an application of
transfer learning that is highly sensitive to personal privacy.
Under black-box and realistic settings, we show that existing inference
techniques are ineffective, as interacting with individual training instances
through the student models does not reveal information about the teacher. We
then propose novel strategies to infer from aggregate-level information.
Consequently, membership inference attacks on the teacher model are shown to be
possible, even when the adversary has access only to the student models.
We further demonstrate that sensitive attributes can be inferred, even in the
case where the adversary has limited auxiliary information. Finally, defensive
strategies are discussed and evaluated. Our extensive study indicates that
information leakage is a real privacy threat to the transfer learning framework
widely used in real-life situations.
Related papers
- Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models [112.48136829374741]
In this paper, we unveil a new vulnerability: the privacy backdoor attack.
When a victim fine-tunes a backdoored model, their training data will be leaked at a significantly higher rate than if they had fine-tuned a typical model.
Our findings highlight a critical privacy concern within the machine learning community and call for a reevaluation of safety protocols in the use of open-source pre-trained models.
arXiv Detail & Related papers (2024-04-01T16:50:54Z) - Segue: Side-information Guided Generative Unlearnable Examples for
Facial Privacy Protection in Real World [64.4289385463226]
We propose Segue: Side-information guided generative unlearnable examples.
To improve transferability, we introduce side information such as true labels and pseudo labels.
It can resist JPEG compression, adversarial training, and some standard data augmentations.
arXiv Detail & Related papers (2023-10-24T06:22:37Z) - Students Parrot Their Teachers: Membership Inference on Model
Distillation [54.392069096234074]
We study the privacy provided by knowledge distillation to both the teacher and student training sets.
Our attacks are strongest when student and teacher sets are similar, or when the attacker can poison the teacher set.
arXiv Detail & Related papers (2023-03-06T19:16:23Z) - Does CLIP Know My Face? [31.21910897081894]
We introduce a novel method to assess privacy for multi-modal models, specifically vision-language models like CLIP.
The proposed Identity Inference Attack (IDIA) reveals whether an individual was included in the training data by querying the model with images of the same person.
Our results highlight the need for stronger privacy protection in large-scale models and suggest that IDIAs can be used to prove the unauthorized use of data for training and to enforce privacy laws.
arXiv Detail & Related papers (2022-09-15T14:48:50Z) - Teacher Model Fingerprinting Attacks Against Transfer Learning [23.224444604615123]
We present the first comprehensive investigation of the teacher model exposure threat in the transfer learning context.
We propose a teacher model fingerprinting attack to infer the origin of a student model it transfers from.
We show that our attack can accurately identify the model origin with few probing queries.
arXiv Detail & Related papers (2021-06-23T15:52:35Z) - Understanding Robustness in Teacher-Student Setting: A New Perspective [42.746182547068265]
Adrial examples are machine learning models where bounded adversarial perturbation could mislead the models to make arbitrarily incorrect predictions.
Extensive studies try to explain the existence of adversarial examples and provide ways to improve model robustness.
Our studies could shed light on the future exploration about adversarial examples, and enhancing model robustness via principled data augmentation.
arXiv Detail & Related papers (2021-02-25T20:54:24Z) - Sampling Attacks: Amplification of Membership Inference Attacks by
Repeated Queries [74.59376038272661]
We introduce sampling attack, a novel membership inference technique that unlike other standard membership adversaries is able to work under severe restriction of no access to scores of the victim model.
We show that a victim model that only publishes the labels is still susceptible to sampling attacks and the adversary can recover up to 100% of its performance.
For defense, we choose differential privacy in the form of gradient perturbation during the training of the victim model as well as output perturbation at prediction time.
arXiv Detail & Related papers (2020-09-01T12:54:54Z) - Stylized Adversarial Defense [105.88250594033053]
adversarial training creates perturbation patterns and includes them in the training set to robustify the model.
We propose to exploit additional information from the feature space to craft stronger adversaries.
Our adversarial training approach demonstrates strong robustness compared to state-of-the-art defenses.
arXiv Detail & Related papers (2020-07-29T08:38:10Z) - Differentially Private Deep Learning with Smooth Sensitivity [144.31324628007403]
We study privacy concerns through the lens of differential privacy.
In this framework, privacy guarantees are generally obtained by perturbing models in such a way that specifics of data used to train the model are made ambiguous.
One of the most important techniques used in previous works involves an ensemble of teacher models, which return information to a student based on a noisy voting procedure.
In this work, we propose a novel voting mechanism with smooth sensitivity, which we call Immutable Noisy ArgMax, that, under certain conditions, can bear very large random noising from the teacher without affecting the useful information transferred to the student
arXiv Detail & Related papers (2020-03-01T15:38:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.