Towards Multi-Agent Reinforcement Learning driven Over-The-Counter
Market Simulations
- URL: http://arxiv.org/abs/2210.07184v2
- Date: Tue, 1 Aug 2023 15:22:05 GMT
- Title: Towards Multi-Agent Reinforcement Learning driven Over-The-Counter
Market Simulations
- Authors: Nelson Vadori, Leo Ardon, Sumitra Ganesh, Thomas Spooner, Selim
Amrouni, Jared Vann, Mengda Xu, Zeyu Zheng, Tucker Balch, Manuela Veloso
- Abstract summary: We study a game between liquidity provider and liquidity taker agents interacting in an over-the-counter market.
By playing against each other, our deep-reinforcement-learning-driven agents learn emergent behaviors.
We show convergence rates for our multi-agent policy gradient algorithm under a transitivity assumption.
- Score: 16.48389671789281
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study a game between liquidity provider and liquidity taker agents
interacting in an over-the-counter market, for which the typical example is
foreign exchange. We show how a suitable design of parameterized families of
reward functions coupled with shared policy learning constitutes an efficient
solution to this problem. By playing against each other, our
deep-reinforcement-learning-driven agents learn emergent behaviors relative to
a wide spectrum of objectives encompassing profit-and-loss, optimal execution
and market share. In particular, we find that liquidity providers naturally
learn to balance hedging and skewing, where skewing refers to setting their buy
and sell prices asymmetrically as a function of their inventory. We further
introduce a novel RL-based calibration algorithm which we found performed well
at imposing constraints on the game equilibrium. On the theoretical side, we
are able to show convergence rates for our multi-agent policy gradient
algorithm under a transitivity assumption, closely related to generalized
ordinal potential games.
Related papers
- Deviations from the Nash equilibrium and emergence of tacit collusion in a two-player optimal execution game with reinforcement learning [0.9208007322096533]
We study a scenario in which two autonomous agents learn to liquidate the same asset optimally in the presence of market impact.
Our results show that the strategies learned by the agents deviate significantly from the Nash equilibrium of the corresponding market impact game.
We explore how different levels of market volatility influence the agents' performance and the equilibria they discover.
arXiv Detail & Related papers (2024-08-21T16:54:53Z) - A Black-box Approach for Non-stationary Multi-agent Reinforcement Learning [53.83345471268163]
We investigate learning the equilibria in non-stationary multi-agent systems.
We show how to test for various types of equilibria by a black-box reduction to single-agent learning.
arXiv Detail & Related papers (2023-06-12T23:48:24Z) - Context-Aware Bayesian Network Actor-Critic Methods for Cooperative
Multi-Agent Reinforcement Learning [7.784991832712813]
We introduce a Bayesian network to inaugurate correlations between agents' action selections in their joint policy.
We develop practical algorithms to learn the context-aware Bayesian network policies.
Empirical results on a range of MARL benchmarks show the benefits of our approach.
arXiv Detail & Related papers (2023-06-02T21:22:27Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - Learn to Match with No Regret: Reinforcement Learning in Markov Matching
Markets [151.03738099494765]
We study a Markov matching market involving a planner and a set of strategic agents on the two sides of the market.
We propose a reinforcement learning framework that integrates optimistic value iteration with maximum weight matching.
We prove that the algorithm achieves sublinear regret.
arXiv Detail & Related papers (2022-03-07T19:51:25Z) - Finding General Equilibria in Many-Agent Economic Simulations Using Deep
Reinforcement Learning [72.23843557783533]
We show that deep reinforcement learning can discover stable solutions that are epsilon-Nash equilibria for a meta-game over agent types.
Our approach is more flexible and does not need unrealistic assumptions, e.g., market clearing.
We demonstrate our approach in real-business-cycle models, a representative family of DGE models, with 100 worker-consumers, 10 firms, and a government who taxes and redistributes.
arXiv Detail & Related papers (2022-01-03T17:00:17Z) - Learning Equilibria in Matching Markets from Bandit Feedback [139.29934476625488]
We develop a framework and algorithms for learning stable market outcomes under uncertainty.
Our work takes a first step toward elucidating when and how stable matchings arise in large, data-driven marketplaces.
arXiv Detail & Related papers (2021-08-19T17:59:28Z) - Optimal Market Making by Reinforcement Learning [0.0]
We apply Reinforcement Learning algorithms to the classic quantitative finance Market Making problem.
We find that the Deep Q-Learning algorithm manages to recover the optimal agent.
arXiv Detail & Related papers (2021-04-08T20:13:21Z) - Strategic bidding in freight transport using deep reinforcement learning [0.0]
This paper presents a multi-agent reinforcement learning algorithm to represent strategic bidding behavior in freight transport markets.
Using this algorithm, we investigate whether feasible market equilibriums arise without any central control or communication between agents.
arXiv Detail & Related papers (2021-02-18T10:17:10Z) - Learning Strategies in Decentralized Matching Markets under Uncertain
Preferences [91.3755431537592]
We study the problem of decision-making in the setting of a scarcity of shared resources when the preferences of agents are unknown a priori.
Our approach is based on the representation of preferences in a reproducing kernel Hilbert space.
We derive optimal strategies that maximize agents' expected payoffs.
arXiv Detail & Related papers (2020-10-29T03:08:22Z) - Calibration of Shared Equilibria in General Sum Partially Observable
Markov Games [15.572157454411533]
We consider a general sum partially observable Markov game where agents of different types share a single policy network.
This paper aims at i) formally understanding equilibria reached by such agents, and ii) matching emergent phenomena of such equilibria to real-world targets.
arXiv Detail & Related papers (2020-06-23T15:14:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.