Effects of Model Misspecification on Bayesian Bandits: Case Studies in
UX Optimization
- URL: http://arxiv.org/abs/2010.04010v1
- Date: Wed, 7 Oct 2020 14:34:28 GMT
- Title: Effects of Model Misspecification on Bayesian Bandits: Case Studies in
UX Optimization
- Authors: Mack Sweeney, Matthew van Adelsberg, Kathryn Laskey, Carlotta
Domeniconi
- Abstract summary: We present a novel formulation as a restless, sleeping bandit with unobserved confounders plus optional stopping.
Case studies show how common misspecifications can lead to sub-optimal rewards.
We also present the first model to exploit cointegration in a restless bandit, demonstrating that finite regret and fast and consistent optional stopping are possible.
- Score: 8.704145252476705
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bayesian bandits using Thompson Sampling have seen increasing success in
recent years. Yet existing value models (of rewards) are misspecified on many
real-world problem. We demonstrate this on the User Experience Optimization
(UXO) problem, providing a novel formulation as a restless, sleeping bandit
with unobserved confounders plus optional stopping. Our case studies show how
common misspecifications can lead to sub-optimal rewards, and we provide model
extensions to address these, along with a scientific model building process
practitioners can adopt or adapt to solve their own unique problems. To our
knowledge, this is the first study showing the effects of overdispersion on
bandit explore/exploit efficacy, tying the common notions of under- and
over-confidence to over- and under-exploration, respectively. We also present
the first model to exploit cointegration in a restless bandit, demonstrating
that finite regret and fast and consistent optional stopping are possible by
moving beyond simpler windowing, discounting, and drift models.
Related papers
- Neural Dueling Bandits [58.90189511247936]
We use a neural network to estimate the reward function using preference feedback for the previously selected arms.
We then extend our theoretical results to contextual bandit problems with binary feedback, which is in itself a non-trivial contribution.
arXiv Detail & Related papers (2024-07-24T09:23:22Z) - Causal Bandits: The Pareto Optimal Frontier of Adaptivity, a Reduction to Linear Bandits, and Limitations around Unknown Marginals [28.94461817548213]
We prove upper and matching lower bounds on the possible trade-offs in the performance of learning in conditionally benign and arbitrary environments.
We are the first to obtain instance-dependent bounds for causal bandits, by reducing the problem to the linear bandit setting.
arXiv Detail & Related papers (2024-07-01T04:12:15Z) - Predictive Churn with the Set of Good Models [64.05949860750235]
We study the effect of conflicting predictions over the set of near-optimal machine learning models.
We present theoretical results on the expected churn between models within the Rashomon set.
We show how our approach can be used to better anticipate, reduce, and avoid churn in consumer-facing applications.
arXiv Detail & Related papers (2024-02-12T16:15:25Z) - Source-Free Unsupervised Domain Adaptation with Hypothesis Consolidation
of Prediction Rationale [53.152460508207184]
Source-Free Unsupervised Domain Adaptation (SFUDA) is a challenging task where a model needs to be adapted to a new domain without access to target domain labels or source domain data.
This paper proposes a novel approach that considers multiple prediction hypotheses for each sample and investigates the rationale behind each hypothesis.
To achieve the optimal performance, we propose a three-step adaptation process: model pre-adaptation, hypothesis consolidation, and semi-supervised learning.
arXiv Detail & Related papers (2024-02-02T05:53:22Z) - Dropout-Based Rashomon Set Exploration for Efficient Predictive
Multiplicity Estimation [15.556756363296543]
Predictive multiplicity refers to the phenomenon in which classification tasks admit multiple competing models that achieve almost-equally-optimal performance.
We propose a novel framework that utilizes dropout techniques for exploring models in the Rashomon set.
We show that our technique consistently outperforms baselines in terms of the effectiveness of predictive multiplicity metric estimation.
arXiv Detail & Related papers (2024-02-01T16:25:00Z) - Single-Model Attribution of Generative Models Through Final-Layer Inversion [16.506531590300806]
We propose a new approach for single-model attribution in the open-world setting based on final-layer inversion and anomaly detection.
We show that the utilized final-layer inversion can be reduced to a convex lasso optimization problem, making our approach theoretically sound and computationally efficient.
arXiv Detail & Related papers (2023-05-26T13:06:38Z) - Making Substitute Models More Bayesian Can Enhance Transferability of
Adversarial Examples [89.85593878754571]
transferability of adversarial examples across deep neural networks is the crux of many black-box attacks.
We advocate to attack a Bayesian model for achieving desirable transferability.
Our method outperforms recent state-of-the-arts by large margins.
arXiv Detail & Related papers (2023-02-10T07:08:13Z) - Bias-Robust Bayesian Optimization via Dueling Bandit [57.82422045437126]
We consider Bayesian optimization in settings where observations can be adversarially biased.
We propose a novel approach for dueling bandits based on information-directed sampling (IDS)
Thereby, we obtain the first efficient kernelized algorithm for dueling bandits that comes with cumulative regret guarantees.
arXiv Detail & Related papers (2021-05-25T10:08:41Z) - A Twin Neural Model for Uplift [59.38563723706796]
Uplift is a particular case of conditional treatment effect modeling.
We propose a new loss function defined by leveraging a connection with the Bayesian interpretation of the relative risk.
We show our proposed method is competitive with the state-of-the-art in simulation setting and on real data from large scale randomized experiments.
arXiv Detail & Related papers (2021-05-11T16:02:39Z) - Towards Robust and Reliable Algorithmic Recourse [11.887537452826624]
We propose a novel framework, RObust Algorithmic Recourse (ROAR), that leverages adversarial training for finding recourses that are robust to model shifts.
We also carry out detailed theoretical analysis which underscores the importance of constructing recourses that are robust to model shifts.
arXiv Detail & Related papers (2021-02-26T17:38:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.