Learning Equilibria in Matching Markets from Bandit Feedback
- URL: http://arxiv.org/abs/2108.08843v1
- Date: Thu, 19 Aug 2021 17:59:28 GMT
- Title: Learning Equilibria in Matching Markets from Bandit Feedback
- Authors: Meena Jagadeesan, Alexander Wei, Yixin Wang, Michael I. Jordan, Jacob
Steinhardt
- Abstract summary: We develop a framework and algorithms for learning stable market outcomes under uncertainty.
Our work takes a first step toward elucidating when and how stable matchings arise in large, data-driven marketplaces.
- Score: 139.29934476625488
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large-scale, two-sided matching platforms must find market outcomes that
align with user preferences while simultaneously learning these preferences
from data. However, since preferences are inherently uncertain during learning,
the classical notion of stability (Gale and Shapley, 1962; Shapley and Shubik,
1971) is unattainable in these settings. To bridge this gap, we develop a
framework and algorithms for learning stable market outcomes under uncertainty.
Our primary setting is matching with transferable utilities, where the platform
both matches agents and sets monetary transfers between them. We design an
incentive-aware learning objective that captures the distance of a market
outcome from equilibrium. Using this objective, we analyze the complexity of
learning as a function of preference structure, casting learning as a
stochastic multi-armed bandit problem. Algorithmically, we show that "optimism
in the face of uncertainty," the principle underlying many bandit algorithms,
applies to a primal-dual formulation of matching with transfers and leads to
near-optimal regret bounds. Our work takes a first step toward elucidating when
and how stable matchings arise in large, data-driven marketplaces.
Related papers
- Putting Gale & Shapley to Work: Guaranteeing Stability Through Learning [14.448192914855674]
Two-sided matching markets describe a class of problems wherein participants from one side of the market must be matched to those from the other side according to their preferences.
We exploit the structure of stable solutions to devise algorithms that improve the likelihood of finding stable solutions.
arXiv Detail & Related papers (2024-10-06T06:47:53Z) - Deviations from the Nash equilibrium and emergence of tacit collusion in a two-player optimal execution game with reinforcement learning [0.9208007322096533]
We study a scenario in which two autonomous agents learn to liquidate the same asset optimally in the presence of market impact.
Our results show that the strategies learned by the agents deviate significantly from the Nash equilibrium of the corresponding market impact game.
We explore how different levels of market volatility influence the agents' performance and the equilibria they discover.
arXiv Detail & Related papers (2024-08-21T16:54:53Z) - Decentralized and Uncoordinated Learning of Stable Matchings: A Game-Theoretic Approach [9.376820789668304]
We consider the problem of learning stable matchings with unknown preferences in a decentralized and uncoordinated manner.
We show that applying the exponential weight (EXP) learning algorithm to the stable matching game achieves logarithmic regret in a fully decentralized and uncoordinated fashion.
We also introduce another decentralized and uncoordinated learning algorithm that globally converges to a stable matching with arbitrarily high probability.
arXiv Detail & Related papers (2024-07-31T02:36:14Z) - Fairness in Matching under Uncertainty [78.39459690570531]
algorithmic two-sided marketplaces have drawn attention to the issue of fairness in such settings.
We axiomatize a notion of individual fairness in the two-sided marketplace setting which respects the uncertainty in the merits.
We design a linear programming framework to find fair utility-maximizing distributions over allocations.
arXiv Detail & Related papers (2023-02-08T00:30:32Z) - Decentralized, Communication- and Coordination-free Learning in
Structured Matching Markets [2.9833943723592764]
We study the problem of online learning in competitive settings in the context of two-sided matching markets.
We propose a class of decentralized, communication- and coordination-free algorithms that agents can use to reach to their stable match.
arXiv Detail & Related papers (2022-06-06T04:08:04Z) - Learn to Match with No Regret: Reinforcement Learning in Markov Matching
Markets [151.03738099494765]
We study a Markov matching market involving a planner and a set of strategic agents on the two sides of the market.
We propose a reinforcement learning framework that integrates optimistic value iteration with maximum weight matching.
We prove that the algorithm achieves sublinear regret.
arXiv Detail & Related papers (2022-03-07T19:51:25Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Multi-Stage Decentralized Matching Markets: Uncertain Preferences and
Strategic Behaviors [91.3755431537592]
This article develops a framework for learning optimal strategies in real-world matching markets.
We show that there exists a welfare-versus-fairness trade-off that is characterized by the uncertainty level of acceptance.
We prove that participants can be better off with multi-stage matching compared to single-stage matching.
arXiv Detail & Related papers (2021-02-13T19:25:52Z) - Learning Strategies in Decentralized Matching Markets under Uncertain
Preferences [91.3755431537592]
We study the problem of decision-making in the setting of a scarcity of shared resources when the preferences of agents are unknown a priori.
Our approach is based on the representation of preferences in a reproducing kernel Hilbert space.
We derive optimal strategies that maximize agents' expected payoffs.
arXiv Detail & Related papers (2020-10-29T03:08:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.