On the Efficacy of Differentially Private Few-shot Image Classification
- URL: http://arxiv.org/abs/2302.01190v3
- Date: Tue, 19 Dec 2023 09:45:36 GMT
- Title: On the Efficacy of Differentially Private Few-shot Image Classification
- Authors: Marlon Tobaben, Aliaksandra Shysheya, John Bronskill, Andrew Paverd,
Shruti Tople, Santiago Zanella-Beguelin, Richard E Turner, Antti Honkela
- Abstract summary: In many applications including personalization and federated learning, it is crucial to perform well in the few-shot setting.
We show how the accuracy and vulnerability to attack of few-shot DP image classification models are affected as the number of shots per class, privacy level, model architecture, downstream dataset, and subset of learnable parameters in the model vary.
- Score: 40.49270725252068
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There has been significant recent progress in training differentially private
(DP) models which achieve accuracy that approaches the best non-private models.
These DP models are typically pretrained on large public datasets and then
fine-tuned on private downstream datasets that are relatively large and similar
in distribution to the pretraining data. However, in many applications
including personalization and federated learning, it is crucial to perform well
(i) in the few-shot setting, as obtaining large amounts of labeled data may be
problematic; and (ii) on datasets from a wide variety of domains for use in
various specialist settings. To understand under which conditions few-shot DP
can be effective, we perform an exhaustive set of experiments that reveals how
the accuracy and vulnerability to attack of few-shot DP image classification
models are affected as the number of shots per class, privacy level, model
architecture, downstream dataset, and subset of learnable parameters in the
model vary. We show that to achieve DP accuracy on par with non-private models,
the shots per class must be increased as the privacy level increases. We also
show that learning parameter-efficient FiLM adapters under DP is competitive
with learning just the final classifier layer or learning all of the network
parameters. Finally, we evaluate DP federated learning systems and establish
state-of-the-art performance on the challenging FLAIR benchmark.
Related papers
- Differentially Private Fine-Tuning of Diffusion Models [22.454127503937883]
The integration of Differential Privacy with diffusion models (DMs) presents a promising yet challenging frontier.
Recent developments in this field have highlighted the potential for generating high-quality synthetic data by pre-training on public data.
We propose a strategy optimized for private diffusion models, which minimizes the number of trainable parameters to enhance the privacy-utility trade-off.
arXiv Detail & Related papers (2024-06-03T14:18:04Z) - Pre-training Differentially Private Models with Limited Public Data [54.943023722114134]
differential privacy (DP) is a prominent method to gauge the degree of security provided to the models.
DP is yet not capable of protecting a substantial portion of the data used during the initial pre-training stage.
We develop a novel DP continual pre-training strategy using only 10% of public data.
Our strategy can achieve DP accuracy of 41.5% on ImageNet-21k, as well as non-DP accuracy of 55.7% and and 60.0% on downstream tasks Places365 and iNaturalist-2021.
arXiv Detail & Related papers (2024-02-28T23:26:27Z) - Differentially Private Diffusion Models [46.46256537222917]
We build on the recent success of diffusion models (DMs) and introduce Differentially Private Diffusion Models (DPDMs)
We propose noise multiplicity, a powerful modification of DP-SGD tailored to the training of DMs.
We validate our novel DPDMs on image generation benchmarks and achieve state-of-the-art performance in all experiments.
arXiv Detail & Related papers (2022-10-18T15:20:47Z) - Large Scale Transfer Learning for Differentially Private Image
Classification [51.10365553035979]
Differential Privacy (DP) provides a formal framework for training machine learning models with individual example level privacy.
Private training using DP-SGD protects against leakage by injecting noise into individual example gradients.
While this result is quite appealing, the computational cost of training large-scale models with DP-SGD is substantially higher than non-private training.
arXiv Detail & Related papers (2022-05-06T01:22:20Z) - Large Language Models Can Be Strong Differentially Private Learners [70.0317718115406]
Differentially Private (DP) learning has seen limited success for building large deep learning models of text.
We show that this performance drop can be mitigated with the use of large pretrained models.
We propose a memory saving technique that allows clipping in DP-SGD to run without instantiating per-example gradients.
arXiv Detail & Related papers (2021-10-12T01:45:27Z) - An Efficient DP-SGD Mechanism for Large Scale NLP Models [28.180412581994485]
Data used to train Natural Language Understanding (NLU) models may contain private information such as addresses or phone numbers.
It is desirable that underlying models do not expose private information contained in the training data.
Differentially Private Gradient Descent (DP-SGD) has been proposed as a mechanism to build privacy-preserving models.
arXiv Detail & Related papers (2021-07-14T15:23:27Z) - DPlis: Boosting Utility of Differentially Private Deep Learning via
Randomized Smoothing [0.0]
We propose DPlis--Differentially Private Learning wIth Smoothing.
We show that DPlis can effectively boost model quality and training stability under a given privacy budget.
arXiv Detail & Related papers (2021-03-02T06:33:14Z) - Knowledge-Enriched Distributional Model Inversion Attacks [49.43828150561947]
Model inversion (MI) attacks are aimed at reconstructing training data from model parameters.
We present a novel inversion-specific GAN that can better distill knowledge useful for performing attacks on private models from public data.
Our experiments show that the combination of these techniques can significantly boost the success rate of the state-of-the-art MI attacks by 150%.
arXiv Detail & Related papers (2020-10-08T16:20:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.