Coordinated Pandemic Control with Large Language Model Agents as Policymaking Assistants
- URL: http://arxiv.org/abs/2601.09264v1
- Date: Wed, 14 Jan 2026 07:59:44 GMT
- Title: Coordinated Pandemic Control with Large Language Model Agents as Policymaking Assistants
- Authors: Ziyi Shi, Xusen Guo, Hongliang Lu, Mingxing Peng, Haotian Wang, Zheng Zhu, Zhenning Li, Yuxuan Liang, Xinhu Zheng, Hai Yang,
- Abstract summary: We propose a large language model (LLM) multi-agent policymaking framework that supports coordinated and proactive pandemic control across regions.<n>By integrating real-world data, a pandemic evolution simulator, and structured inter-agent communication, our framework enables agents to jointly explore counterfactual intervention scenarios.<n>Compared with real-world pandemic outcomes, our approach reduces cumulative infections and deaths by up to 63.7% and 40.1%, respectively, at the individual state level.
- Score: 51.26321657927398
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Effective pandemic control requires timely and coordinated policymaking across administrative regions that are intrinsically interdependent. However, human-driven responses are often fragmented and reactive, with policies formulated in isolation and adjusted only after outbreaks escalate, undermining proactive intervention and global pandemic mitigation. To address this challenge, here we propose a large language model (LLM) multi-agent policymaking framework that supports coordinated and proactive pandemic control across regions. Within our framework, each administrative region is assigned an LLM agent as an AI policymaking assistant. The agent reasons over region-specific epidemiological dynamics while communicating with other agents to account for cross-regional interdependencies. By integrating real-world data, a pandemic evolution simulator, and structured inter-agent communication, our framework enables agents to jointly explore counterfactual intervention scenarios and synthesize coordinated policy decisions through a closed-loop simulation process. We validate the proposed framework using state-level COVID-19 data from the United States between April and December 2020, together with real-world mobility records and observed policy interventions. Compared with real-world pandemic outcomes, our approach reduces cumulative infections and deaths by up to 63.7% and 40.1%, respectively, at the individual state level, and by 39.0% and 27.0%, respectively, when aggregated across states. These results demonstrate that LLM multi-agent systems can enable more effective pandemic control with coordinated policymaking...
Related papers
- LLM-Powered Social Digital Twins: A Framework for Simulating Population Behavioral Response to Policy Interventions [0.2787288702904897]
Social Digital Twins are virtual population replicas where Large Language Models serve as cognitive engines for individual agents.<n>We instantiate this framework in the domain of pandemic response, using COVID-19 as a case study.<n>We discuss implications for policy simulation, limitations of the approach, and directions for extending LLM-based digital twins beyond pandemic response.
arXiv Detail & Related papers (2026-01-03T13:25:33Z) - DispatchMAS: Fusing taxonomy and artificial intelligence agents for emergency medical services [49.70819009392778]
Large Language Models (LLMs) and Multi-Agent Systems (MAS) offer opportunities to augment dispatchers.<n>This study aimed to develop and evaluate a taxonomy-grounded, multi-agent system for simulating realistic scenarios.
arXiv Detail & Related papers (2025-10-24T08:01:21Z) - Learning Pareto-Optimal Pandemic Intervention Policies with MORL [1.160208922584163]
We develop a framework for modeling and evaluating disease-spread prevention strategies.<n>Our simulator reproduces national-scale pandemic dynamics with orders of magnitude higher fidelity than other models.<n>This work supports transparent, evidence-based policymaking for mitigating public health crises.
arXiv Detail & Related papers (2025-10-02T12:06:29Z) - MF-LLM: Simulating Population Decision Dynamics via a Mean-Field Large Language Model Framework [53.82097200295448]
Mean-Field LLM (MF-LLM) is first to incorporate mean field theory into social simulation.<n>MF-LLM models bidirectional interactions between individuals and the population through an iterative process.<n> IB-Tune is a novel fine-tuning method inspired by the Information Bottleneck principle.
arXiv Detail & Related papers (2025-04-30T12:41:51Z) - Federated Natural Policy Gradient and Actor Critic Methods for Multi-task Reinforcement Learning [46.28771270378047]
Federated reinforcement learning (RL) enables collaborative decision making of multiple distributed agents without sharing local data trajectories.
In this work, we consider a multi-task setting, in which each agent has its own private reward function corresponding to different tasks, while sharing the same transition kernel of the environment.
We learn a globally optimal policy that maximizes the sum of the discounted total rewards of all the agents in a decentralized manner.
arXiv Detail & Related papers (2023-11-01T00:15:18Z) - Evaluation of non-pharmaceutical interventions and optimal strategies
for containing the COVID-19 pandemic [14.807368322926227]
We investigate associations between policies, mobility patterns, and virus transmission.
Results highlight the power of state of emergency declaration and wearing face masks.
Our framework can be extended to inform policy makers of any country about best practices in pandemic response.
arXiv Detail & Related papers (2022-02-28T17:33:25Z) - An Agent-Based Model of COVID-19 Diffusion to Plan and Evaluate
Intervention Policies [0.09236074230806579]
The model includes the structural data of Piedmont, an Italian region.
The model is generative of complex epidemic dynamics emerging from the consequences of agents' actions and interactions.
arXiv Detail & Related papers (2021-08-19T19:23:17Z) - Optimal Epidemic Control as a Contextual Combinatorial Bandit with
Budget [26.49683079770031]
In light of the COVID-19 pandemic, it is an open challenge and critical practical problem to find a optimal way to prescribe the best policies.
To solve this multi-dimensional tradeoff of exploitation and exploration, we formulate this technical challenge as a contextual bandit problem.
Agent should generate useful intervention plans that policy makers can implement in real time to minimize both the number of daily COVID-19 cases and the stringency of the recommended interventions.
arXiv Detail & Related papers (2021-06-30T04:46:31Z) - Dealing with Non-Stationarity in Multi-Agent Reinforcement Learning via
Trust Region Decomposition [52.06086375833474]
Non-stationarity is one thorny issue in multi-agent reinforcement learning.
We introduce a $delta$-stationarity measurement to explicitly model the stationarity of a policy sequence.
We propose a trust region decomposition network based on message passing to estimate the joint policy divergence.
arXiv Detail & Related papers (2021-02-21T14:46:50Z) - When and How to Lift the Lockdown? Global COVID-19 Scenario Analysis and
Policy Assessment using Compartmental Gaussian Processes [111.69190108272133]
coronavirus disease 2019 (COVID-19) global pandemic has led many countries to impose unprecedented lockdown measures.
Data-driven models that predict COVID-19 fatalities under different lockdown policy scenarios are essential.
This paper develops a Bayesian model for predicting the effects of COVID-19 lockdown policies in a global context.
arXiv Detail & Related papers (2020-05-13T18:21:50Z) - Hierarchical Reinforcement Learning for Automatic Disease Diagnosis [52.111516253474285]
We propose to integrate a hierarchical policy structure of two levels into the dialogue systemfor policy learning.
The proposed policy structure is capable to deal with diagnosis problem including large number of diseases and symptoms.
arXiv Detail & Related papers (2020-04-29T15:02:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.