Information-Gathering in Latent Bandits
- URL: http://arxiv.org/abs/2207.03635v1
- Date: Fri, 8 Jul 2022 01:15:12 GMT
- Title: Information-Gathering in Latent Bandits
- Authors: Alexander Galozy, Slawomir Nowaczyk
- Abstract summary: We present a method for information-gathering in latent bandits.
We show that choosing the best arm given the agent's beliefs about the states incurs higher regret.
We also show that by choosing arms carefully, we obtain an improved estimation of the state distribution.
- Score: 79.6953033727455
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the latent bandit problem, the learner has access to reward distributions
and -- for the non-stationary variant -- transition models of the environment.
The reward distributions are conditioned on the arm and unknown latent states.
The goal is to use the reward history to identify the latent state, allowing
for the optimal choice of arms in the future. The latent bandit setting lends
itself to many practical applications, such as recommender and decision support
systems, where rich data allows the offline estimation of environment models
with online learning remaining a critical component. Previous solutions in this
setting always choose the highest reward arm according to the agent's beliefs
about the state, not explicitly considering the value of information-gathering
arms. Such information-gathering arms do not necessarily provide the highest
reward, thus may never be chosen by an agent that chooses the highest reward
arms at all times.
In this paper, we present a method for information-gathering in latent
bandits. Given particular reward structures and transition matrices, we show
that choosing the best arm given the agent's beliefs about the states incurs
higher regret. Furthermore, we show that by choosing arms carefully, we obtain
an improved estimation of the state distribution, and thus lower the cumulative
regret through better arm choices in the future. We evaluate our method on both
synthetic and real-world data sets, showing significant improvement in regret
over state-of-the-art methods.
Related papers
- Optimal Best Arm Identification with Fixed Confidence in Restless Bandits [66.700654953613]
We study best arm identification in a restless multi-armed bandit setting with finitely many arms.
The discrete-time data generated by each arm forms a homogeneous Markov chain taking values in a common, finite state space.
It is demonstrated that tracking the long-term behavior of a certain Markov decision process and its state-action visitation proportions are the key ingredients in analyzing the converse and achievability bounds.
arXiv Detail & Related papers (2023-10-20T10:04:05Z) - Pure Exploration under Mediators' Feedback [63.56002444692792]
Multi-armed bandits are a sequential-decision-making framework, where, at each interaction step, the learner selects an arm and observes a reward.
We consider the scenario in which the learner has access to a set of mediators, each of which selects the arms on the agent's behalf according to a and possibly unknown policy.
We propose a sequential decision-making strategy for discovering the best arm under the assumption that the mediators' policies are known to the learner.
arXiv Detail & Related papers (2023-08-29T18:18:21Z) - Covariance Adaptive Best Arm Identification [0.0]
The goal is to identify the arm with the highest mean reward with a probability of at least 1 -- $delta$, while minimizing the number of arm pulls.
We propose a more flexible scenario where arms can be dependent and rewards can be sampled simultaneously.
This framework is relevant in various applications, such as clinical trials, where similarities between patients or drugs suggest underlying correlations.
arXiv Detail & Related papers (2023-06-05T06:57:09Z) - Repeated Principal-Agent Games with Unobserved Agent Rewards and
Perfect-Knowledge Agents [5.773269033551628]
We study a scenario of repeated principal-agent games within a multi-armed bandit (MAB) framework.
We design our policy by first constructing an estimator for the agent's expected reward for each bandit arm.
We conclude with numerical simulations demonstrating the applicability of our policy to real-life setting from collaborative transportation planning.
arXiv Detail & Related papers (2023-04-14T21:57:16Z) - Incentive-Aware Recommender Systems in Two-Sided Markets [49.692453629365204]
We propose a novel recommender system that aligns with agents' incentives while achieving myopically optimal performance.
Our framework models this incentive-aware system as a multi-agent bandit problem in two-sided markets.
Both algorithms satisfy an ex-post fairness criterion, which protects agents from over-exploitation.
arXiv Detail & Related papers (2022-11-23T22:20:12Z) - Best Arm Identification in Restless Markov Multi-Armed Bandits [85.55466536537293]
We study the problem of identifying the best arm in a multi-armed bandit environment.
A decision entity wishes to find the index of the best arm as quickly as possible, subject to an upper bound error probability.
We show that this policy achieves an upper bound that depends on $R$ and is monotonically non-increasing as $Rtoinfty$.
arXiv Detail & Related papers (2022-03-29T04:58:04Z) - The Role of Contextual Information in Best Arm Identification [13.651941268805693]
We study the best-arm identification problem with fixed confidence when contextual information is available in bandits.
We show the instance-specific sample complexity lower bounds for the problem.
We experimentally confirm that context information contributes to faster best-arm identification.
arXiv Detail & Related papers (2021-06-26T18:39:38Z) - Robustness Guarantees for Mode Estimation with an Application to Bandits [131.21717367564963]
We introduce a theory for multi-armed bandits where the values are the modes of the reward distributions instead of the mean.
We show in simulations that our algorithms are robust to perturbation of the arms by adversarial noise sequences.
arXiv Detail & Related papers (2020-03-05T21:29:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.