Reinforcement Learning of Sequential Price Mechanisms
- URL: http://arxiv.org/abs/2010.01180v2
- Date: Wed, 5 May 2021 20:01:05 GMT
- Title: Reinforcement Learning of Sequential Price Mechanisms
- Authors: Gianluca Brero, Alon Eden, Matthias Gerstgrasser, David C. Parkes,
Duncan Rheingans-Yoo
- Abstract summary: We introduce the use of reinforcement learning for indirect mechanisms, working with the existing class of sequential price mechanisms.
We show that our approach can learn optimal or near-optimal mechanisms in several experimental settings.
- Score: 24.302600030585275
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce the use of reinforcement learning for indirect mechanisms,
working with the existing class of sequential price mechanisms, which
generalizes both serial dictatorship and posted price mechanisms and
essentially characterizes all strongly obviously strategyproof mechanisms.
Learning an optimal mechanism within this class forms a partially-observable
Markov decision process. We provide rigorous conditions for when this class of
mechanisms is more powerful than simpler static mechanisms, for sufficiency or
insufficiency of observation statistics for learning, and for the necessity of
complex (deep) policies. We show that our approach can learn optimal or
near-optimal mechanisms in several experimental settings.
Related papers
- Compete and Compose: Learning Independent Mechanisms for Modular World Models [57.94106862271727]
We present COMET, a modular world model which leverages reusable, independent mechanisms across different environments.
COMET is trained on multiple environments with varying dynamics via a two-step process: competition and composition.
We show that COMET is able to adapt to new environments with varying numbers of objects with improved sample efficiency compared to more conventional finetuning approaches.
arXiv Detail & Related papers (2024-04-23T15:03:37Z) - Deep Generative Model-based Synthesis of Four-bar Linkage Mechanisms
with Target Conditions [22.164394511786874]
We propose a deep learning-based generative model for generating multiple crank-rocker four-bar linkage mechanisms.
The proposed model is based on a conditional generative adversarial network (cGAN) with modifications for mechanism synthesis.
The results demonstrate that the proposed model successfully generates multiple distinct mechanisms that satisfy specific kinematic and quasi-static requirements.
arXiv Detail & Related papers (2024-02-22T03:31:00Z) - Competition of Mechanisms: Tracing How Language Models Handle Facts and Counterfactuals [82.68757839524677]
Interpretability research aims to bridge the gap between empirical success and our scientific understanding of large language models (LLMs)
We propose a formulation of competition of mechanisms, which focuses on the interplay of multiple mechanisms instead of individual mechanisms.
Our findings show traces of the mechanisms and their competition across various model components and reveal attention positions that effectively control the strength of certain mechanisms.
arXiv Detail & Related papers (2024-02-18T17:26:51Z) - Deep Learning Meets Mechanism Design: Key Results and Some Novel
Applications [1.2661010067882734]
We present, from relevant literature, technical details of using a deep learning approach for mechanism design.
We demonstrate the power of this approach for three illustrative case studies.
arXiv Detail & Related papers (2024-01-11T06:09:32Z) - Refined Mechanism Design for Approximately Structured Priors via Active
Regression [50.71772232237571]
We consider the problem of a revenue-maximizing seller with a large number of items for sale to $n$ strategic bidders.
It is well-known that optimal and even approximately-optimal mechanisms for this setting are notoriously difficult to characterize or compute.
arXiv Detail & Related papers (2023-10-11T20:34:17Z) - Pessimism meets VCG: Learning Dynamic Mechanism Design via Offline
Reinforcement Learning [114.36124979578896]
We design a dynamic mechanism using offline reinforcement learning algorithms.
Our algorithm is based on the pessimism principle and only requires a mild assumption on the coverage of the offline data set.
arXiv Detail & Related papers (2022-05-05T05:44:26Z) - Properties from Mechanisms: An Equivariance Perspective on Identifiable
Representation Learning [79.4957965474334]
Key goal of unsupervised representation learning is "inverting" a data generating process to recover its latent properties.
This paper asks, "Can we instead identify latent properties by leveraging knowledge of the mechanisms that govern their evolution?"
We provide a complete characterization of the sources of non-identifiability as we vary knowledge about a set of possible mechanisms.
arXiv Detail & Related papers (2021-10-29T14:04:08Z) - Learning Revenue-Maximizing Auctions With Differentiable Matching [50.62088223117716]
We propose a new architecture to approximately learn incentive compatible, revenue-maximizing auctions from sampled valuations.
Our architecture uses the Sinkhorn algorithm to perform a differentiable bipartite matching which allows the network to learn strategyproof revenue-maximizing mechanisms.
arXiv Detail & Related papers (2021-06-15T04:37:57Z) - Near Instance-Optimality in Differential Privacy [38.8726789833284]
We develop notions of instance optimality in differential privacy inspired by classical statistical theory.
We also develop inverse sensitivity mechanisms, which are instance optimal (or nearly instance optimal) for a large class of estimands.
arXiv Detail & Related papers (2020-05-16T04:53:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.