Fine-Tuning with Differential Privacy Necessitates an Additional
Hyperparameter Search
- URL: http://arxiv.org/abs/2210.02156v1
- Date: Wed, 5 Oct 2022 11:32:49 GMT
- Title: Fine-Tuning with Differential Privacy Necessitates an Additional
Hyperparameter Search
- Authors: Yannis Cattan, Christopher A. Choquette-Choo, Nicolas Papernot,
Abhradeep Thakurta
- Abstract summary: We show how carefully selecting the layers being fine-tuned in the pretrained neural network allows us to establish new state-of-the-art tradeoffs between privacy and accuracy.
We achieve 77.9% accuracy for $(varepsilon, delta)= (2, 10-5)$ on CIFAR-100 for a model pretrained on ImageNet.
- Score: 38.83524780461911
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Models need to be trained with privacy-preserving learning algorithms to
prevent leakage of possibly sensitive information contained in their training
data. However, canonical algorithms like differentially private stochastic
gradient descent (DP-SGD) do not benefit from model scale in the same way as
non-private learning. This manifests itself in the form of unappealing
tradeoffs between privacy and utility (accuracy) when using DP-SGD on complex
tasks. To remediate this tension, a paradigm is emerging: fine-tuning with
differential privacy from a model pretrained on public (i.e., non-sensitive)
training data.
In this work, we identify an oversight of existing approaches for
differentially private fine tuning. They do not tailor the fine-tuning approach
to the specifics of learning with privacy. Our main result is to show how
carefully selecting the layers being fine-tuned in the pretrained neural
network allows us to establish new state-of-the-art tradeoffs between privacy
and accuracy. For instance, we achieve 77.9% accuracy for $(\varepsilon,
\delta)=(2, 10^{-5})$ on CIFAR-100 for a model pretrained on ImageNet. Our work
calls for additional hyperparameter search to configure the differentially
private fine-tuning procedure itself.
Related papers
- TAN Without a Burn: Scaling Laws of DP-SGD [70.7364032297978]
Differentially Private methods for training Deep Neural Networks (DNNs) have progressed recently.
We decouple privacy analysis and experimental behavior of noisy training to explore the trade-off with minimal computational requirements.
We apply the proposed method on CIFAR-10 and ImageNet and, in particular, strongly improve the state-of-the-art on ImageNet with a +9 points gain in top-1 accuracy.
arXiv Detail & Related papers (2022-10-07T08:44:35Z) - Individual Privacy Accounting for Differentially Private Stochastic Gradient Descent [69.14164921515949]
We characterize privacy guarantees for individual examples when releasing models trained by DP-SGD.
We find that most examples enjoy stronger privacy guarantees than the worst-case bound.
This implies groups that are underserved in terms of model utility simultaneously experience weaker privacy guarantees.
arXiv Detail & Related papers (2022-06-06T13:49:37Z) - Large Scale Transfer Learning for Differentially Private Image
Classification [51.10365553035979]
Differential Privacy (DP) provides a formal framework for training machine learning models with individual example level privacy.
Private training using DP-SGD protects against leakage by injecting noise into individual example gradients.
While this result is quite appealing, the computational cost of training large-scale models with DP-SGD is substantially higher than non-private training.
arXiv Detail & Related papers (2022-05-06T01:22:20Z) - Mixed Differential Privacy in Computer Vision [133.68363478737058]
AdaMix is an adaptive differentially private algorithm for training deep neural network classifiers using both private and public image data.
A few-shot or even zero-shot learning baseline that ignores private data can outperform fine-tuning on a large private dataset.
arXiv Detail & Related papers (2022-03-22T06:15:43Z) - Do Not Let Privacy Overbill Utility: Gradient Embedding Perturbation for
Private Learning [74.73901662374921]
A differentially private model degrades the utility drastically when the model comprises a large number of trainable parameters.
We propose an algorithm emphGradient Embedding Perturbation (GEP) towards training differentially private deep models with decent accuracy.
arXiv Detail & Related papers (2021-02-25T04:29:58Z) - Tempered Sigmoid Activations for Deep Learning with Differential Privacy [33.574715000662316]
We show that the choice of activation function is central to bounding the sensitivity of privacy-preserving deep learning.
We achieve new state-of-the-art accuracy on MNIST, FashionMNIST, and CIFAR10 without any modification of the learning procedure fundamentals.
arXiv Detail & Related papers (2020-07-28T13:19:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.