A Blessing of Dimensionality in Membership Inference through
Regularization
- URL: http://arxiv.org/abs/2205.14055v2
- Date: Fri, 14 Apr 2023 02:21:26 GMT
- Title: A Blessing of Dimensionality in Membership Inference through
Regularization
- Authors: Jasper Tan, Daniel LeJeune, Blake Mason, Hamid Javadi, Richard G.
Baraniuk
- Abstract summary: We show how the number of parameters of a model can induce a privacy--utility trade-off.
We then show that if coupled with proper generalization regularization, increasing the number of parameters of a model can actually increase both its privacy and performance.
- Score: 29.08230123469755
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Is overparameterization a privacy liability? In this work, we study the
effect that the number of parameters has on a classifier's vulnerability to
membership inference attacks. We first demonstrate how the number of parameters
of a model can induce a privacy--utility trade-off: increasing the number of
parameters generally improves generalization performance at the expense of
lower privacy. However, remarkably, we then show that if coupled with proper
regularization, increasing the number of parameters of a model can actually
simultaneously increase both its privacy and performance, thereby eliminating
the privacy--utility trade-off. Theoretically, we demonstrate this curious
phenomenon for logistic regression with ridge regularization in a bi-level
feature ensemble setting. Pursuant to our theoretical exploration, we develop a
novel leave-one-out analysis tool to precisely characterize the vulnerability
of a linear classifier to the optimal membership inference attack. We
empirically exhibit this "blessing of dimensionality" for neural networks on a
variety of tasks using early stopping as the regularizer.
Related papers
- Pseudo-Probability Unlearning: Towards Efficient and Privacy-Preserving Machine Unlearning [59.29849532966454]
We propose PseudoProbability Unlearning (PPU), a novel method that enables models to forget data to adhere to privacy-preserving manner.
Our method achieves over 20% improvements in forgetting error compared to the state-of-the-art.
arXiv Detail & Related papers (2024-11-04T21:27:06Z) - Defending Membership Inference Attacks via Privacy-aware Sparsity Tuning [9.39508697970812]
We propose a simple fix to the L1 Regularization, by employing adaptive penalties to different parameters.
Our key idea behind PAST is to promote sparsity in parameters that significantly contribute to privacy leakage.
Using PAST, the network shrinks the loss gap between members and non-members, leading to strong resistance to privacy attacks.
arXiv Detail & Related papers (2024-10-09T12:13:49Z) - Initialization Matters: Privacy-Utility Analysis of Overparameterized
Neural Networks [72.51255282371805]
We prove a privacy bound for the KL divergence between model distributions on worst-case neighboring datasets.
We find that this KL privacy bound is largely determined by the expected squared gradient norm relative to model parameters during training.
arXiv Detail & Related papers (2023-10-31T16:13:22Z) - Theoretically Principled Federated Learning for Balancing Privacy and
Utility [61.03993520243198]
We propose a general learning framework for the protection mechanisms that protects privacy via distorting model parameters.
It can achieve personalized utility-privacy trade-off for each model parameter, on each client, at each communication round in federated learning.
arXiv Detail & Related papers (2023-05-24T13:44:02Z) - Revisiting Hyperparameter Tuning with Differential Privacy [1.6425841685973384]
We provide a framework for privacy-preserving machine learning with differential privacy.
We show that its additional privacy loss bound incurred by hyperparameter tuning is upper-bounded by the squared root of the gained utility.
We note that the additional privacy loss bound would empirically scale like a squared root of the logarithm of the utility term, benefiting from the design of doubling step.
arXiv Detail & Related papers (2022-11-03T14:42:19Z) - Monotonicity and Double Descent in Uncertainty Estimation with Gaussian
Processes [52.92110730286403]
It is commonly believed that the marginal likelihood should be reminiscent of cross-validation metrics and that both should deteriorate with larger input dimensions.
We prove that by tuning hyper parameters, the performance, as measured by the marginal likelihood, improves monotonically with the input dimension.
We also prove that cross-validation metrics exhibit qualitatively different behavior that is characteristic of double descent.
arXiv Detail & Related papers (2022-10-14T08:09:33Z) - Provably tuning the ElasticNet across instances [53.0518090093538]
We consider the problem of tuning the regularization parameters of Ridge regression, LASSO, and the ElasticNet across multiple problem instances.
Our results are the first general learning-theoretic guarantees for this important class of problems.
arXiv Detail & Related papers (2022-07-20T21:22:40Z) - A Differentially Private Framework for Deep Learning with Convexified
Loss Functions [4.059849656394191]
Differential privacy (DP) has been applied in deep learning for preserving privacy of the underlying training sets.
Existing DP practice falls into three categories - objective perturbation, gradient perturbation and output perturbation.
We propose a novel output perturbation framework by injecting DP noise into a randomly sampled neuron.
arXiv Detail & Related papers (2022-04-03T11:10:05Z) - Parameters or Privacy: A Provable Tradeoff Between Overparameterization
and Membership Inference [29.743945643424553]
Over parameterized models generalize well (small error on the test data) even when trained to memorize the training data (zero error on the training data)
This has led to an arms race towards increasingly over parameterized models (c.f., deep learning)
arXiv Detail & Related papers (2022-02-02T19:00:21Z) - Robustness Threats of Differential Privacy [70.818129585404]
We experimentally demonstrate that networks, trained with differential privacy, in some settings might be even more vulnerable in comparison to non-private versions.
We study how the main ingredients of differentially private neural networks training, such as gradient clipping and noise addition, affect the robustness of the model.
arXiv Detail & Related papers (2020-12-14T18:59:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.