Pessimism meets VCG: Learning Dynamic Mechanism Design via Offline
Reinforcement Learning
- URL: http://arxiv.org/abs/2205.02450v1
- Date: Thu, 5 May 2022 05:44:26 GMT
- Title: Pessimism meets VCG: Learning Dynamic Mechanism Design via Offline
Reinforcement Learning
- Authors: Boxiang Lyu and Zhaoran Wang and Mladen Kolar and Zhuoran Yang
- Abstract summary: We design a dynamic mechanism using offline reinforcement learning algorithms.
Our algorithm is based on the pessimism principle and only requires a mild assumption on the coverage of the offline data set.
- Score: 114.36124979578896
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Dynamic mechanism design has garnered significant attention from both
computer scientists and economists in recent years. By allowing agents to
interact with the seller over multiple rounds, where agents' reward functions
may change with time and are state dependent, the framework is able to model a
rich class of real world problems. In these works, the interaction between
agents and sellers are often assumed to follow a Markov Decision Process (MDP).
We focus on the setting where the reward and transition functions of such an
MDP are not known a priori, and we are attempting to recover the optimal
mechanism using an a priori collected data set. In the setting where the
function approximation is employed to handle large state spaces, with only mild
assumptions on the expressiveness of the function class, we are able to design
a dynamic mechanism using offline reinforcement learning algorithms. Moreover,
learned mechanisms approximately have three key desiderata: efficiency,
individual rationality, and truthfulness. Our algorithm is based on the
pessimism principle and only requires a mild assumption on the coverage of the
offline data set. To the best of our knowledge, our work provides the first
offline RL algorithm for dynamic mechanism design without assuming uniform
coverage.
Related papers
- Nash Incentive-compatible Online Mechanism Learning via Weakly Differentially Private Online Learning [6.869373893556194]
We study a multi-round mechanism design problem, where we interact with a set of agents over a sequence of rounds.
We wish to design an incentive-compatible (IC) online learning scheme to maximize an application-specific objective.
arXiv Detail & Related papers (2024-07-06T00:02:25Z) - Refined Mechanism Design for Approximately Structured Priors via Active
Regression [50.71772232237571]
We consider the problem of a revenue-maximizing seller with a large number of items for sale to $n$ strategic bidders.
It is well-known that optimal and even approximately-optimal mechanisms for this setting are notoriously difficult to characterize or compute.
arXiv Detail & Related papers (2023-10-11T20:34:17Z) - Multi-Agent Reinforcement Learning for Long-Term Network Resource
Allocation through Auction: a V2X Application [7.326507804995567]
We formulate offloading of computational tasks from a dynamic group of mobile agents (e.g., cars) as decentralized decision making among autonomous agents.
We design an interaction mechanism that incentivizes such agents to align private and system goals by balancing between competition and cooperation.
We propose a novel multi-agent online learning algorithm that learns with partial, delayed and noisy state information.
arXiv Detail & Related papers (2022-07-29T10:29:06Z) - Stabilizing Q-learning with Linear Architectures for Provably Efficient
Learning [53.17258888552998]
This work proposes an exploration variant of the basic $Q$-learning protocol with linear function approximation.
We show that the performance of the algorithm degrades very gracefully under a novel and more permissive notion of approximation error.
arXiv Detail & Related papers (2022-06-01T23:26:51Z) - Learning Dynamic Mechanisms in Unknown Environments: A Reinforcement
Learning Approach [130.9259586568977]
We propose novel learning algorithms to recover the dynamic Vickrey-Clarke-Grove (VCG) mechanism over multiple rounds of interaction.
A key contribution of our approach is incorporating reward-free online Reinforcement Learning (RL) to aid exploration over a rich policy space.
arXiv Detail & Related papers (2022-02-25T16:17:23Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.