Computing the optimal distributionally-robust strategy to commit to
- URL: http://arxiv.org/abs/2209.07647v1
- Date: Thu, 15 Sep 2022 23:20:26 GMT
- Title: Computing the optimal distributionally-robust strategy to commit to
- Authors: Sai Mali Ananthanarayanan and Christian Kroer
- Abstract summary: We show that a distributionally-robust Stackelberg equilibrium always exists across a wide array of uncertainty models.
We present two algorithms to compute a distributionally-robust strong Stackelberg equilibrium.
Experiments substantiate the tractability of our algorithm on a classical Stackelberg game.
- Score: 32.1464237233989
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The Stackelberg game model, where a leader commits to a strategy and the
follower best responds, has found widespread application, particularly to
security problems. In the security setting, the goal is for the leader to
compute an optimal strategy to commit to, in order to protect some asset. In
many of these applications, the parameters of the follower utility model are
not known with certainty. Distributionally-robust optimization addresses this
issue by allowing a distribution over possible model parameters, where this
distribution comes from a set of possible distributions. The goal is to
maximize the expected utility with respect to the worst-case distribution. We
initiate the study of distributionally-robust models for computing the optimal
strategy to commit to. We consider the case of normal-form games with
uncertainty about the follower utility model. Our main theoretical result is to
show that a distributionally-robust Stackelberg equilibrium always exists
across a wide array of uncertainty models. For the case of a finite set of
possible follower utility functions we present two algorithms to compute a
distributionally-robust strong Stackelberg equilibrium (DRSSE) using
mathematical programs. Next, in the general case where there is an infinite
number of possible follower utility functions and the uncertainty is
represented by a Wasserstein ball around a finitely-supported nominal
distribution, we give an incremental mixed-integer-programming-based algorithm
for computing the optimal distributionally-robust strategy. Experiments
substantiate the tractability of our algorithm on a classical Stackelberg game,
showing that our approach scales to medium-sized games.
Related papers
- Distributionally Robust Skeleton Learning of Discrete Bayesian Networks [9.46389554092506]
We consider the problem of learning the exact skeleton of general discrete Bayesian networks from potentially corrupted data.
We propose to optimize the most adverse risk over a family of distributions within bounded Wasserstein distance or KL divergence to the empirical distribution.
We present efficient algorithms and show the proposed methods are closely related to the standard regularized regression approach.
arXiv Detail & Related papers (2023-11-10T15:33:19Z) - Generalized Schrödinger Bridge Matching [54.171931505066]
Generalized Schr"odinger Bridge (GSB) problem setup is prevalent in many scientific areas both within and without machine learning.
We propose Generalized Schr"odinger Bridge Matching (GSBM), a new matching algorithm inspired by recent advances.
We show that such a generalization can be cast as solving conditional optimal control, for which variational approximations can be used.
arXiv Detail & Related papers (2023-10-03T17:42:11Z) - Value-Distributional Model-Based Reinforcement Learning [59.758009422067]
Quantifying uncertainty about a policy's long-term performance is important to solve sequential decision-making tasks.
We study the problem from a model-based Bayesian reinforcement learning perspective.
We propose Epistemic Quantile-Regression (EQR), a model-based algorithm that learns a value distribution function.
arXiv Detail & Related papers (2023-08-12T14:59:19Z) - Distributional Hamilton-Jacobi-Bellman Equations for Continuous-Time
Reinforcement Learning [39.07307690074323]
We consider the problem of predicting the distribution of returns obtained by an agent interacting in a continuous-time environment.
Accurate return predictions have proven useful for determining optimal policies for risk-sensitive control, state representations, multiagent coordination, and more.
We propose a tractable algorithm for approximately solving the distributional HJB based on a JKO scheme, which can be implemented in an online control algorithm.
arXiv Detail & Related papers (2022-05-24T16:33:54Z) - Wrapped Distributions on homogeneous Riemannian manifolds [58.720142291102135]
Control over distributions' properties, such as parameters, symmetry and modality yield a family of flexible distributions.
We empirically validate our approach by utilizing our proposed distributions within a variational autoencoder and a latent space network model.
arXiv Detail & Related papers (2022-04-20T21:25:21Z) - Implicit Distributional Reinforcement Learning [61.166030238490634]
implicit distributional actor-critic (IDAC) built on two deep generator networks (DGNs)
Semi-implicit actor (SIA) powered by a flexible policy distribution.
We observe IDAC outperforms state-of-the-art algorithms on representative OpenAI Gym environments.
arXiv Detail & Related papers (2020-07-13T02:52:18Z) - A Convergent and Dimension-Independent Min-Max Optimization Algorithm [32.492526162436405]
We show that a distribution which the min-player uses to update its parameters depends on a smooth and bounded nonnon-concave function.
Our algorithm converges to an approximate local equilibrium in a number of iterations that does not depend on the iterations.
arXiv Detail & Related papers (2020-06-22T16:11:30Z) - A Distributional Analysis of Sampling-Based Reinforcement Learning
Algorithms [67.67377846416106]
We present a distributional approach to theoretical analyses of reinforcement learning algorithms for constant step-sizes.
We show that value-based methods such as TD($lambda$) and $Q$-Learning have update rules which are contractive in the space of distributions of functions.
arXiv Detail & Related papers (2020-03-27T05:13:29Z) - Distributionally Robust Bayesian Quadrature Optimization [60.383252534861136]
We study BQO under distributional uncertainty in which the underlying probability distribution is unknown except for a limited set of its i.i.d. samples.
A standard BQO approach maximizes the Monte Carlo estimate of the true expected objective given the fixed sample set.
We propose a novel posterior sampling based algorithm, namely distributionally robust BQO (DRBQO) for this purpose.
arXiv Detail & Related papers (2020-01-19T12:00:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.