Dynamic Global Sensitivity for Differentially Private Contextual Bandits
- URL: http://arxiv.org/abs/2208.14555v1
- Date: Tue, 30 Aug 2022 22:09:08 GMT
- Title: Dynamic Global Sensitivity for Differentially Private Contextual Bandits
- Authors: Huazheng Wang, David Zhao, Hongning Wang
- Abstract summary: We propose a differentially private linear contextual bandit algorithm, via a tree-based mechanism to add Laplace or Gaussian noise to model parameters.
Our key insight is that as the model converges during online update, the global sensitivity of its parameters shrinks over time.
We provide a rigorous theoretical analysis over the amount of noise added via dynamic global sensitivity and the corresponding upper regret bound of our proposed algorithm.
- Score: 37.413361483987835
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bandit algorithms have become a reference solution for interactive
recommendation. However, as such algorithms directly interact with users for
improved recommendations, serious privacy concerns have been raised regarding
its practical use. In this work, we propose a differentially private linear
contextual bandit algorithm, via a tree-based mechanism to add Laplace or
Gaussian noise to model parameters. Our key insight is that as the model
converges during online update, the global sensitivity of its parameters
shrinks over time (thus named dynamic global sensitivity). Compared with
existing solutions, our dynamic global sensitivity analysis allows us to inject
less noise to obtain $(\epsilon, \delta)$-differential privacy with added
regret caused by noise injection in $\tilde O(\log{T}\sqrt{T}/\epsilon)$. We
provide a rigorous theoretical analysis over the amount of noise added via
dynamic global sensitivity and the corresponding upper regret bound of our
proposed algorithm. Experimental results on both synthetic and real-world
datasets confirmed the algorithm's advantage against existing solutions.
Related papers
- Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - Contextual Dynamic Pricing: Algorithms, Optimality, and Local Differential Privacy Constraints [10.057344315478709]
We study contextual dynamic pricing problems where a firm sells products to $T$ sequentially-arriving consumers.
We first show the optimal regret is of order $sqrtdT$, up to logarithmic factors.
We extend our study to dynamic pricing under mixed privacy constraints, improving the privacy-utility tradeoff by leveraging public data.
arXiv Detail & Related papers (2024-06-04T15:44:10Z) - Information limits and Thouless-Anderson-Palmer equations for spiked matrix models with structured noise [19.496063739638924]
We consider a saturate problem of Bayesian inference for a structured spiked model.
We show how to predict the statistical limits using an efficient algorithm inspired by the theory of adaptive Thouless-Anderson-Palmer equations.
arXiv Detail & Related papers (2024-05-31T16:38:35Z) - Geometry of Sensitivity: Twice Sampling and Hybrid Clipping in Differential Privacy with Optimal Gaussian Noise and Application to Deep Learning [18.92302645198466]
We study the problem of the construction of optimal randomization in Differential Privacy.
Finding the minimal perturbation for properly-selected sensitivity sets is a central problem in DP research.
arXiv Detail & Related papers (2023-09-06T02:45:08Z) - On the Complexity of Differentially Private Best-Arm Identification with
Fixed Confidence [16.295693624977563]
We study the problem of Best Arm Identification with fixed confidence under $epsilon$-global Differential Privacy.
Our lower bound suggests the existence of two privacy regimes depending on the privacy budget.
We propose AdaP-TT, an $epsilon$-global DP variant of the Top Two algorithm.
arXiv Detail & Related papers (2023-09-05T13:07:25Z) - Adaptive Noise Covariance Estimation under Colored Noise using Dynamic
Expectation Maximization [1.550120821358415]
We introduce a novel brain-inspired algorithm that accurately estimates the NCM for dynamic systems subjected to colored noise.
We mathematically prove that our NCM estimator converges to the global optimum of this free energy objective.
Notably, we show that our method outperforms the best baseline (Variational Bayes) in joint noise and state estimation for high colored noise.
arXiv Detail & Related papers (2023-08-15T14:21:53Z) - Optimistic Active Exploration of Dynamical Systems [52.91573056896633]
We develop an algorithm for active exploration called OPAX.
We show how OPAX can be reduced to an optimal control problem that can be solved at each episode.
Our experiments show that OPAX is not only theoretically sound but also performs well for zero-shot planning on novel downstream tasks.
arXiv Detail & Related papers (2023-06-21T16:26:59Z) - Online Continuous Hyperparameter Optimization for Generalized Linear Contextual Bandits [55.03293214439741]
In contextual bandits, an agent sequentially makes actions from a time-dependent action set based on past experience.
We propose the first online continuous hyperparameter tuning framework for contextual bandits.
We show that it could achieve a sublinear regret in theory and performs consistently better than all existing methods on both synthetic and real datasets.
arXiv Detail & Related papers (2023-02-18T23:31:20Z) - Normalized/Clipped SGD with Perturbation for Differentially Private
Non-Convex Optimization [94.06564567766475]
DP-SGD and DP-NSGD mitigate the risk of large models memorizing sensitive training data.
We show that these two algorithms achieve similar best accuracy while DP-NSGD is comparatively easier to tune than DP-SGD.
arXiv Detail & Related papers (2022-06-27T03:45:02Z) - Greedy versus Map-based Optimized Adaptive Algorithms for
random-telegraph-noise mitigation by spectator qubits [6.305016513788048]
In a scenario where data-storage qubits are kept in isolation as far as possible, noise mitigation can still be done using additional noise probes.
We construct a theoretical model assuming projective measurements on the qubits, and derive the performance of different measurement and control strategies.
We show, analytically and numerically, that MOAAAR outperforms the Greedy algorithm, especially in the regime of high noise sensitivity of SQ.
arXiv Detail & Related papers (2022-05-25T08:25:10Z) - Optimizing Information-theoretical Generalization Bounds via Anisotropic
Noise in SGLD [73.55632827932101]
We optimize the information-theoretical generalization bound by manipulating the noise structure in SGLD.
We prove that with constraint to guarantee low empirical risk, the optimal noise covariance is the square root of the expected gradient covariance.
arXiv Detail & Related papers (2021-10-26T15:02:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.