Finding General Equilibria in Many-Agent Economic Simulations Using Deep
Reinforcement Learning
- URL: http://arxiv.org/abs/2201.01163v1
- Date: Mon, 3 Jan 2022 17:00:17 GMT
- Title: Finding General Equilibria in Many-Agent Economic Simulations Using Deep
Reinforcement Learning
- Authors: Michael Curry, Alexander Trott, Soham Phade, Yu Bai, Stephan Zheng
- Abstract summary: We show that deep reinforcement learning can discover stable solutions that are epsilon-Nash equilibria for a meta-game over agent types.
Our approach is more flexible and does not need unrealistic assumptions, e.g., market clearing.
We demonstrate our approach in real-business-cycle models, a representative family of DGE models, with 100 worker-consumers, 10 firms, and a government who taxes and redistributes.
- Score: 72.23843557783533
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real economies can be seen as a sequential imperfect-information game with
many heterogeneous, interacting strategic agents of various agent types, such
as consumers, firms, and governments. Dynamic general equilibrium models are
common economic tools to model the economic activity, interactions, and
outcomes in such systems. However, existing analytical and computational
methods struggle to find explicit equilibria when all agents are strategic and
interact, while joint learning is unstable and challenging. Amongst others, a
key reason is that the actions of one economic agent may change the reward
function of another agent, e.g., a consumer's expendable income changes when
firms change prices or governments change taxes. We show that multi-agent deep
reinforcement learning (RL) can discover stable solutions that are epsilon-Nash
equilibria for a meta-game over agent types, in economic simulations with many
agents, through the use of structured learning curricula and efficient GPU-only
simulation and training. Conceptually, our approach is more flexible and does
not need unrealistic assumptions, e.g., market clearing, that are commonly used
for analytical tractability. Our GPU implementation enables training and
analyzing economies with a large number of agents within reasonable time
frames, e.g., training completes within a day. We demonstrate our approach in
real-business-cycle models, a representative family of DGE models, with 100
worker-consumers, 10 firms, and a government who taxes and redistributes. We
validate the learned meta-game epsilon-Nash equilibria through approximate
best-response analyses, show that RL policies align with economic intuitions,
and that our approach is constructive, e.g., by explicitly learning a spectrum
of meta-game epsilon-Nash equilibria in open RBC models.
Related papers
- EconAgent: Large Language Model-Empowered Agents for Simulating Macroeconomic Activities [43.70290385026672]
We introduce EconAgent, a large language model-empowered agent with human-like characteristics for macroeconomic simulation.
We first construct a simulation environment that incorporates various market dynamics driven by agents' decisions.
Through the perception module, we create heterogeneous agents with distinct decision-making mechanisms.
arXiv Detail & Related papers (2023-10-16T14:19:40Z) - MERMAIDE: Learning to Align Learners using Model-Based Meta-Learning [62.065503126104126]
We study how a principal can efficiently and effectively intervene on the rewards of a previously unseen learning agent in order to induce desirable outcomes.
This is relevant to many real-world settings like auctions or taxation, where the principal may not know the learning behavior nor the rewards of real people.
We introduce MERMAIDE, a model-based meta-learning framework to train a principal that can quickly adapt to out-of-distribution agents.
arXiv Detail & Related papers (2023-04-10T15:44:50Z) - Many learning agents interacting with an agent-based market model [0.0]
We consider the dynamics of learning optimal execution trading agents interacting with a reactive Agent-Based Model.
The model represents a market ecology with 3-trophic levels represented by: optimal execution learning agents, minimally intelligent liquidity takers, and fast electronic liquidity providers.
We examine whether the inclusion of optimal execution agents that can learn is able to produce dynamics with the same complexity as empirical data.
arXiv Detail & Related papers (2023-03-13T18:15:52Z) - Towards Multi-Agent Reinforcement Learning driven Over-The-Counter
Market Simulations [16.48389671789281]
We study a game between liquidity provider and liquidity taker agents interacting in an over-the-counter market.
By playing against each other, our deep-reinforcement-learning-driven agents learn emergent behaviors.
We show convergence rates for our multi-agent policy gradient algorithm under a transitivity assumption.
arXiv Detail & Related papers (2022-10-13T17:06:08Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - Modeling Bounded Rationality in Multi-Agent Simulations Using Rationally
Inattentive Reinforcement Learning [85.86440477005523]
We study more human-like RL agents which incorporate an established model of human-irrationality, the Rational Inattention (RI) model.
RIRL models the cost of cognitive information processing using mutual information.
We show that using RIRL yields a rich spectrum of new equilibrium behaviors that differ from those found under rational assumptions.
arXiv Detail & Related papers (2022-01-18T20:54:00Z) - ERMAS: Becoming Robust to Reward Function Sim-to-Real Gaps in
Multi-Agent Simulations [110.72725220033983]
Epsilon-Robust Multi-Agent Simulation (ERMAS) is a framework for learning AI policies that are robust to such multiagent sim-to-real gaps.
ERMAS learns tax policies that are robust to changes in agent risk aversion, improving social welfare by up to 15% in complextemporal simulations.
In particular, ERMAS learns tax policies that are robust to changes in agent risk aversion, improving social welfare by up to 15% in complextemporal simulations.
arXiv Detail & Related papers (2021-06-10T04:32:20Z) - Decentralized Reinforcement Learning: Global Decision-Making via Local
Economic Transactions [80.49176924360499]
We establish a framework for directing a society of simple, specialized, self-interested agents to solve sequential decision problems.
We derive a class of decentralized reinforcement learning algorithms.
We demonstrate the potential advantages of a society's inherent modular structure for more efficient transfer learning.
arXiv Detail & Related papers (2020-07-05T16:41:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.