Policy-focused Agent-based Modeling using RL Behavioral Models
- URL: http://arxiv.org/abs/2006.05048v3
- Date: Thu, 5 Nov 2020 20:41:17 GMT
- Title: Policy-focused Agent-based Modeling using RL Behavioral Models
- Authors: Osonde A. Osoba, Raffaele Vardavas, Justin Grana, Rushil Zutshi, Amber
Jaycocks
- Abstract summary: This paper examines the value of reinforcement learning models as adaptive, high-performing, and behaviorally-valid models of agent decision-making in ABMs.
We test the hypothesis that RL agents are effective as utility-maximizing agents in policy ABMs.
Experiments show that RL behavioral models are effective at producing reward-seeking or reward-maximizing behaviors in ABM agents.
- Score: 0.40498500266986387
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Agent-based Models (ABMs) are valuable tools for policy analysis. ABMs help
analysts explore the emergent consequences of policy interventions in
multi-agent decision-making settings. But the validity of inferences drawn from
ABM explorations depends on the quality of the ABM agents' behavioral models.
Standard specifications of agent behavioral models rely either on heuristic
decision-making rules or on regressions trained on past data. Both prior
specification modes have limitations. This paper examines the value of
reinforcement learning (RL) models as adaptive, high-performing, and
behaviorally-valid models of agent decision-making in ABMs. We test the
hypothesis that RL agents are effective as utility-maximizing agents in policy
ABMs. We also address the problem of adapting RL algorithms to handle
multi-agency in games by adapting and extending methods from recent literature.
We evaluate the performance of such RL-based ABM agents via experiments on two
policy-relevant ABMs: a minority game ABM, and an ABM of Influenza
Transmission. We run some analytic experiments on our AI-equipped ABMs e.g.
explorations of the effects of behavioral heterogeneity in a population and the
emergence of synchronization in a population. The experiments show that RL
behavioral models are effective at producing reward-seeking or
reward-maximizing behaviors in ABM agents. Furthermore, RL behavioral models
can learn to outperform the default adaptive behavioral models in the two ABMs
examined.
Related papers
- Domain-driven Metrics for Reinforcement Learning: A Case Study on Epidemic Control using Agent-based Simulation [0.29360071145551064]
In this study, we are developing domain-driven metrics for RL, while building on state-of-the-art metrics.<n>Results show the use of domain-driven rewards in conjunction with traditional and state-of-the-art metrics for a few different simulation scenarios.
arXiv Detail & Related papers (2025-08-07T08:40:19Z) - Learning Individual Behavior in Agent-Based Models with Graph Diffusion Networks [2.749593964424624]
Agent-Based Models (ABMs) are powerful tools for studying emergent properties in complex systems.<n>We propose a novel framework to learn a differentiable surrogate of any ABM by observing its generated data.<n>Our method combines diffusion models to capture behaviorality and graph neural networks to model agent interactions.
arXiv Detail & Related papers (2025-05-27T16:55:56Z) - Echo Chamber: RL Post-training Amplifies Behaviors Learned in Pretraining [74.83412846804977]
Reinforcement learning (RL)-based fine-tuning has become a crucial step in post-training language models.
We present a systematic end-to-end study of RL fine-tuning for mathematical reasoning by training models entirely from scratch.
arXiv Detail & Related papers (2025-04-10T17:15:53Z) - Reinforcement Learning for Machine Learning Model Deployment: Evaluating Multi-Armed Bandits in ML Ops Environments [0.0]
We investigate whether reinforcement learning (RL)-based model management can manage deployment decisions more effectively.
Our approach enables more adaptive production environments by continuously evaluating deployed models and rolling back underperforming ones in real-time.
Our findings suggest that RL-based model management can improve automation, reduce reliance on manual interventions, and mitigate risks associated with post-deployment model failures.
arXiv Detail & Related papers (2025-03-28T16:42:21Z) - On the limits of agency in agent-based models [13.130587222524305]
Agent-based modeling offers powerful insights into complex systems, but its practical utility has been limited by computational constraints.
Recent advancements in large language models (LLMs) could enhance ABMs with adaptive agents, but their integration into large-scale simulations remains challenging.
We present LLM archetypes, a technique that balances behavioral complexity with computational efficiency, allowing for nuanced agent behavior in large-scale simulations.
arXiv Detail & Related papers (2024-09-14T04:17:24Z) - Model-Based RL for Mean-Field Games is not Statistically Harder than Single-Agent RL [57.745700271150454]
We study the sample complexity of reinforcement learning in Mean-Field Games (MFGs) with model-based function approximation.
We introduce the Partial Model-Based Eluder Dimension (P-MBED), a more effective notion to characterize the model class complexity.
arXiv Detail & Related papers (2024-02-08T14:54:47Z) - Learning and Calibrating Heterogeneous Bounded Rational Market Behaviour
with Multi-Agent Reinforcement Learning [4.40301653518681]
Agent-based models (ABMs) have shown promise for modelling various real world phenomena incompatible with traditional equilibrium analysis.
Recent developments in multi-agent reinforcement learning (MARL) offer a way to address this issue from a rationality perspective.
We propose a novel technique for representing heterogeneous processing-constrained agents within a MARL framework.
arXiv Detail & Related papers (2024-02-01T17:21:45Z) - Predictable MDP Abstraction for Unsupervised Model-Based RL [93.91375268580806]
We propose predictable MDP abstraction (PMA)
Instead of training a predictive model on the original MDP, we train a model on a transformed MDP with a learned action space.
We theoretically analyze PMA and empirically demonstrate that PMA leads to significant improvements over prior unsupervised model-based RL approaches.
arXiv Detail & Related papers (2023-02-08T07:37:51Z) - On learning agent-based models from data [22.387666772159974]
Agent-Based Models (ABMs) are used in several fields to study the evolution of complex systems from micro-level assumptions.
We propose a protocol to learn the latent micro-variables of an ABM from data.
arXiv Detail & Related papers (2022-05-10T17:08:26Z) - Training and Evaluation of Deep Policies using Reinforcement Learning
and Generative Models [67.78935378952146]
GenRL is a framework for solving sequential decision-making problems.
It exploits the combination of reinforcement learning and latent variable generative models.
We experimentally determine the characteristics of generative models that have most influence on the performance of the final policy training.
arXiv Detail & Related papers (2022-04-18T22:02:32Z) - Dynamic Noises of Multi-Agent Environments Can Improve Generalization:
Agent-based Models meets Reinforcement Learning [2.492300648514128]
We study the benefits of reinforcement learning environments based on agent-based models (ABM)
We show that their non-deterministic dynamics can improve the generalization of RL agents.
Numerical simulations demonstrate that the intrinsic noise in the ABM-based dynamics of the SIR model not only improve the average reward but also allow the RL agent to generalize on a wider ranges of epidemic parameters.
arXiv Detail & Related papers (2022-03-26T09:56:30Z) - Modeling Bounded Rationality in Multi-Agent Simulations Using Rationally
Inattentive Reinforcement Learning [85.86440477005523]
We study more human-like RL agents which incorporate an established model of human-irrationality, the Rational Inattention (RI) model.
RIRL models the cost of cognitive information processing using mutual information.
We show that using RIRL yields a rich spectrum of new equilibrium behaviors that differ from those found under rational assumptions.
arXiv Detail & Related papers (2022-01-18T20:54:00Z) - Pessimistic Model Selection for Offline Deep Reinforcement Learning [56.282483586473816]
Deep Reinforcement Learning (DRL) has demonstrated great potentials in solving sequential decision making problems in many applications.
One main barrier is the over-fitting issue that leads to poor generalizability of the policy learned by DRL.
We propose a pessimistic model selection (PMS) approach for offline DRL with a theoretical guarantee.
arXiv Detail & Related papers (2021-11-29T06:29:49Z) - BEDS-Bench: Behavior of EHR-models under Distributional Shift--A
Benchmark [21.040754460129854]
We release BEDS-Bench, a benchmark for quantifying the behavior of ML models over EHR data under OOD settings.
We evaluate several learning algorithms under BEDS-Bench and find that all of them show poor generalization performance under distributional shift in general.
arXiv Detail & Related papers (2021-07-17T05:53:24Z) - On the model-based stochastic value gradient for continuous
reinforcement learning [50.085645237597056]
We show that simple model-based agents can outperform state-of-the-art model-free agents in terms of both sample-efficiency and final reward.
Our findings suggest that model-based policy evaluation deserves closer attention.
arXiv Detail & Related papers (2020-08-28T17:58:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.