DPVIm: Differentially Private Variational Inference Improved
- URL: http://arxiv.org/abs/2210.15961v1
- Date: Fri, 28 Oct 2022 07:41:32 GMT
- Title: DPVIm: Differentially Private Variational Inference Improved
- Authors: Joonas J\"alk\"o, Lukas Prediger, Antti Honkela, and Samuel Kaski
- Abstract summary: Differentially private (DP) release of multidimensional statistics typically considers an aggregate sensitivity.
Different dimensions of that vector might have widely different magnitudes and therefore DP perturbation disproportionately affects the signal across dimensions.
We observe this problem in the gradient release of the DP-SGD algorithm when using it for variational inference (VI)
- Score: 13.761202518891329
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Differentially private (DP) release of multidimensional statistics typically
considers an aggregate sensitivity, e.g. the vector norm of a high-dimensional
vector. However, different dimensions of that vector might have widely
different magnitudes and therefore DP perturbation disproportionately affects
the signal across dimensions. We observe this problem in the gradient release
of the DP-SGD algorithm when using it for variational inference (VI), where it
manifests in poor convergence as well as high variance in outputs for certain
variational parameters, and make the following contributions: (i) We
mathematically isolate the cause for the difference in magnitudes between
gradient parts corresponding to different variational parameters. Using this as
prior knowledge we establish a link between the gradients of the variational
parameters, and propose an efficient while simple fix for the problem to obtain
a less noisy gradient estimator, which we call $\textit{aligned}$ gradients.
This approach allows us to obtain the updates for the covariance parameter of a
Gaussian posterior approximation without a privacy cost. We compare this to
alternative approaches for scaling the gradients using analytically derived
preconditioning, e.g. natural gradients. (ii) We suggest using iterate
averaging over the DP parameter traces recovered during the training, to reduce
the DP-induced noise in parameter estimates at no additional cost in privacy.
Finally, (iii) to accurately capture the additional uncertainty DP introduces
to the model parameters, we infer the DP-induced noise from the parameter
traces and include that in the learned posteriors to make them $\textit{noise
aware}$. We demonstrate the efficacy of our proposed improvements through
various experiments on real data.
Related papers
- A Bayesian Approach Toward Robust Multidimensional Ellipsoid-Specific Fitting [0.0]
This work presents a novel and effective method for fitting multidimensional ellipsoids to scattered data in the contamination of noise and outliers.
We incorporate a uniform prior distribution to constrain the search for primitive parameters within an ellipsoidal domain.
We apply it to a wide range of practical applications such as microscopy cell counting, 3D reconstruction, geometric shape approximation, and magnetometer calibration tasks.
arXiv Detail & Related papers (2024-07-27T14:31:51Z) - Differentially Private Optimization with Sparse Gradients [60.853074897282625]
We study differentially private (DP) optimization problems under sparsity of individual gradients.
Building on this, we obtain pure- and approximate-DP algorithms with almost optimal rates for convex optimization with sparse gradients.
arXiv Detail & Related papers (2024-04-16T20:01:10Z) - Differentially Private SGD Without Clipping Bias: An Error-Feedback Approach [62.000948039914135]
Using Differentially Private Gradient Descent with Gradient Clipping (DPSGD-GC) to ensure Differential Privacy (DP) comes at the cost of model performance degradation.
We propose a new error-feedback (EF) DP algorithm as an alternative to DPSGD-GC.
We establish an algorithm-specific DP analysis for our proposed algorithm, providing privacy guarantees based on R'enyi DP.
arXiv Detail & Related papers (2023-11-24T17:56:44Z) - Normalized/Clipped SGD with Perturbation for Differentially Private
Non-Convex Optimization [94.06564567766475]
DP-SGD and DP-NSGD mitigate the risk of large models memorizing sensitive training data.
We show that these two algorithms achieve similar best accuracy while DP-NSGD is comparatively easier to tune than DP-SGD.
arXiv Detail & Related papers (2022-06-27T03:45:02Z) - MSTGD:A Memory Stochastic sTratified Gradient Descent Method with an
Exponential Convergence Rate [0.0]
This paper designs a novel underlineMemory underlineStochastic sunderlineTratified Gradient Descend(underlineMSTGD) algorithm with an exponential convergence rate.
Specifically, MSTGD uses two strategies for variance reduction: the first strategy is to perform variance reduction according to the proportion p of used historical gradient.
Unlike most other algorithms that claim to achieve an exponential convergence rate, the convergence rate is independent of parameters such as dataset size N, batch size n, etc.
arXiv Detail & Related papers (2022-02-21T01:36:26Z) - Improving Differentially Private SGD via Randomly Sparsified Gradients [31.295035726077366]
Differentially private gradient observation (DP-SGD) has been widely adopted in deep learning to provide rigorously defined privacy bound compression.
We propose an and utilize RS to strengthen communication cost and strengthen privacy bound compression.
arXiv Detail & Related papers (2021-12-01T21:43:34Z) - Differentiable Annealed Importance Sampling and the Perils of Gradient
Noise [68.44523807580438]
Annealed importance sampling (AIS) and related algorithms are highly effective tools for marginal likelihood estimation.
Differentiability is a desirable property as it would admit the possibility of optimizing marginal likelihood as an objective.
We propose a differentiable algorithm by abandoning Metropolis-Hastings steps, which further unlocks mini-batch computation.
arXiv Detail & Related papers (2021-07-21T17:10:14Z) - Private Stochastic Non-Convex Optimization: Adaptive Algorithms and
Tighter Generalization Bounds [72.63031036770425]
We propose differentially private (DP) algorithms for bound non-dimensional optimization.
We demonstrate two popular deep learning methods on the empirical advantages over standard gradient methods.
arXiv Detail & Related papers (2020-06-24T06:01:24Z) - Differentially Private Variational Autoencoders with Term-wise Gradient
Aggregation [12.880889651679094]
We study how to learn variational autoencoders with a variety of divergences under differential privacy constraints.
We propose term-wise DP-SGD that crafts randomized gradients in two different ways tailored to the compositions of the loss terms.
arXiv Detail & Related papers (2020-06-19T16:12:28Z) - Implicit differentiation of Lasso-type models for hyperparameter
optimization [82.73138686390514]
We introduce an efficient implicit differentiation algorithm, without matrix inversion, tailored for Lasso-type problems.
Our approach scales to high-dimensional data by leveraging the sparsity of the solutions.
arXiv Detail & Related papers (2020-02-20T18:43:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.