Homomorphically Encrypted Linear Contextual Bandit
- URL: http://arxiv.org/abs/2103.09927v1
- Date: Wed, 17 Mar 2021 21:49:21 GMT
- Title: Homomorphically Encrypted Linear Contextual Bandit
- Authors: Evrard Garcelon and Vianney Perchet and Matteo Pirotta
- Abstract summary: Contextual bandit is a framework for online learning in sequential decision-making problems.
We introduce a privacy-preserving bandit framework based on asymmetric encryption.
We show that despite the complexity of the setting, it is possible to learn over encrypted data.
- Score: 39.5858373448478
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contextual bandit is a general framework for online learning in sequential
decision-making problems that has found application in a large range of
domains, including recommendation system, online advertising, clinical trials
and many more. A critical aspect of bandit methods is that they require to
observe the contexts -- i.e., individual or group-level data -- and the rewards
in order to solve the sequential problem. The large deployment in industrial
applications has increased interest in methods that preserve the privacy of the
users. In this paper, we introduce a privacy-preserving bandit framework based
on asymmetric encryption. The bandit algorithm only observes encrypted
information (contexts and rewards) and has no ability to decrypt it. Leveraging
homomorphic encryption, we show that despite the complexity of the setting, it
is possible to learn over encrypted data. We introduce an algorithm that
achieves a $\widetilde{O}(d\sqrt{T})$ regret bound in any linear contextual
bandit problem, while keeping data encrypted.
Related papers
- Encrypted system identification as-a-service via reliable encrypted matrix inversion [0.0]
Encrypted computation opens up promising avenues across a plethora of application domains.
In particular, Arithmetic homomorphic encryption is a natural fit for cloud-based computational services.
This paper presents an encrypted system identification service enabled by a reliable encrypted solution to at least squares problems.
arXiv Detail & Related papers (2024-10-27T20:00:04Z) - Feel-Good Thompson Sampling for Contextual Dueling Bandits [49.450050682705026]
We propose a Thompson sampling algorithm, named FGTS.CDB, for linear contextual dueling bandits.
At the core of our algorithm is a new Feel-Good exploration term specifically tailored for dueling bandits.
Our algorithm achieves nearly minimax-optimal regret, i.e., $tildemathcalO(dsqrt T)$, where $d$ is the model dimension and $T$ is the time horizon.
arXiv Detail & Related papers (2024-04-09T04:45:18Z) - A Survey on Property-Preserving Database Encryption Techniques in the Cloud [0.0]
There are concerns about the security and confidentiality of the outsourced data.
The report at hand presents a survey on common encryption techniques used for storing data in relation Cloud database services.
arXiv Detail & Related papers (2023-12-19T11:50:31Z) - Lightweight Public Key Encryption in Post-Quantum Computing Era [0.0]
Confidentiality in our digital world is based on the security of cryptographic algorithms.
In the course of technological progress with quantum computers, the protective function of common encryption algorithms is threatened.
Our concept describes the transformation of a classical asymmetric encryption method to a modern complexity class.
arXiv Detail & Related papers (2023-11-24T21:06:42Z) - Contexts can be Cheap: Solving Stochastic Contextual Bandits with Linear
Bandit Algorithms [39.70492757288025]
We address the contextual linear bandit problem, where a decision maker is provided a context.
We show that the contextual problem can be solved as a linear bandit problem.
Our results imply a $O(dsqrtTlog T)$ high-probability regret bound for contextual linear bandits.
arXiv Detail & Related papers (2022-11-08T22:18:53Z) - Syndicated Bandits: A Framework for Auto Tuning Hyper-parameters in
Contextual Bandit Algorithms [74.55200180156906]
The contextual bandit problem models the trade-off between exploration and exploitation.
We show our Syndicated Bandits framework can achieve the optimal regret upper bounds.
arXiv Detail & Related papers (2021-06-05T22:30:21Z) - Robust Stochastic Linear Contextual Bandits Under Adversarial Attacks [81.13338949407205]
Recent works show that optimal bandit algorithms are vulnerable to adversarial attacks and can fail completely in the presence of attacks.
Existing robust bandit algorithms only work for the non-contextual setting under the attack of rewards.
We provide the first robust bandit algorithm for linear contextual bandit setting under a fully adaptive and omniscient attack.
arXiv Detail & Related papers (2021-06-05T22:20:34Z) - A brief history on Homomorphic learning: A privacy-focused approach to
machine learning [2.055949720959582]
Homomorphic encryption allows running arbitrary operations on encrypted data.
It enables us to run any sophisticated machine learning algorithm without access to the underlying raw data.
It took more than 30 years of collective effort to finally find the answer "yes"
arXiv Detail & Related papers (2020-09-09T21:57:47Z) - Stochastic Bandits with Linear Constraints [69.757694218456]
We study a constrained contextual linear bandit setting, where the goal of the agent is to produce a sequence of policies.
We propose an upper-confidence bound algorithm for this problem, called optimistic pessimistic linear bandit (OPLB)
arXiv Detail & Related papers (2020-06-17T22:32:19Z) - Bandits with Partially Observable Confounded Data [74.04376842070624]
We show that this problem is closely related to a variant of the bandit problem with side information.
We construct a linear bandit algorithm that takes advantage of the projected information, and prove regret bounds.
Our results indicate that confounded offline data can significantly improve online learning algorithms.
arXiv Detail & Related papers (2020-06-11T18:48:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.