Optimal Control Policies to Address the Pandemic Health-Economy Dilemma
- URL: http://arxiv.org/abs/2102.12279v1
- Date: Wed, 24 Feb 2021 13:39:07 GMT
- Title: Optimal Control Policies to Address the Pandemic Health-Economy Dilemma
- Authors: Rohit Salgotra, Thomas Seidelmann, Dominik Fischer, Sanaz Mostaghim,
Amiram Moshaiov
- Abstract summary: Non-pharmaceutical interventions (NPIs) are effective measures to contain a pandemic.
Yet, such control measures commonly have a negative effect on the economy.
We propose a macro-level approach to support resolving this Health-Economy Dilemma (HED)
- Score: 2.5649050169091376
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Non-pharmaceutical interventions (NPIs) are effective measures to contain a
pandemic. Yet, such control measures commonly have a negative effect on the
economy. Here, we propose a macro-level approach to support resolving this
Health-Economy Dilemma (HED). First, an extension to the well-known SEIR model
is suggested which includes an economy model. Second, a bi-objective
optimization problem is defined to study optimal control policies in view of
the HED problem. Next, several multi-objective evolutionary algorithms are
applied to perform a study on the health-economy performance trade-offs that
are inherent to the obtained optimal policies. Finally, the results from the
applied algorithms are compared to select a preferred algorithm for future
studies. As expected, for the proposed models and strategies, a clear conflict
between the health and economy performances is found. Furthermore, the results
suggest that the guided usage of NPIs is preferable as compared to refraining
from employing such strategies at all. This study contributes to pandemic
modeling and simulation by providing a novel concept that elaborates on
integrating economic aspects while exploring the optimal moment to enable NPIs.
Related papers
- Efficient Learning of POMDPs with Known Observation Model in Average-Reward Setting [56.92178753201331]
We propose the Observation-Aware Spectral (OAS) estimation technique, which enables the POMDP parameters to be learned from samples collected using a belief-based policy.
We show the consistency of the OAS procedure, and we prove a regret guarantee of order $mathcalO(sqrtT log(T)$ for the proposed OAS-UCRL algorithm.
arXiv Detail & Related papers (2024-10-02T08:46:34Z) - Provably Mitigating Overoptimization in RLHF: Your SFT Loss is Implicitly an Adversarial Regularizer [52.09480867526656]
We identify the source of misalignment as a form of distributional shift and uncertainty in learning human preferences.
To mitigate overoptimization, we first propose a theoretical algorithm that chooses the best policy for an adversarially chosen reward model.
Using the equivalence between reward models and the corresponding optimal policy, the algorithm features a simple objective that combines a preference optimization loss and a supervised learning loss.
arXiv Detail & Related papers (2024-05-26T05:38:50Z) - Provably Efficient UCB-type Algorithms For Learning Predictive State
Representations [55.00359893021461]
The sequential decision-making problem is statistically learnable if it admits a low-rank structure modeled by predictive state representations (PSRs)
This paper proposes the first known UCB-type approach for PSRs, featuring a novel bonus term that upper bounds the total variation distance between the estimated and true models.
In contrast to existing approaches for PSRs, our UCB-type algorithms enjoy computational tractability, last-iterate guaranteed near-optimal policy, and guaranteed model accuracy.
arXiv Detail & Related papers (2023-07-01T18:35:21Z) - Epidemic Control on a Large-Scale-Agent-Based Epidemiology Model using
Deep Deterministic Policy Gradient [0.7244731714427565]
lockdowns, rapid vaccination programs, school closures, and economic stimulus can have positive or unintended negative consequences.
Current research to model and determine an optimal intervention automatically through round-tripping is limited by the simulation objectives, scale (a few thousand individuals), model types that are not suited for intervention studies, and the number of intervention strategies they can explore (discrete vs continuous).
We address these challenges using a Deep Deterministic Policy Gradient (DDPG) based policy optimization framework on a large-scale (100,000 individual) epidemiological agent-based simulation.
arXiv Detail & Related papers (2023-04-10T09:26:07Z) - Evaluating COVID-19 vaccine allocation policies using Bayesian $m$-top
exploration [53.122045119395594]
We present a novel technique for evaluating vaccine allocation strategies using a multi-armed bandit framework.
$m$-top exploration allows the algorithm to learn $m$ policies for which it expects the highest utility.
We consider the Belgian COVID-19 epidemic using the individual-based model STRIDE, where we learn a set of vaccination policies.
arXiv Detail & Related papers (2023-01-30T12:22:30Z) - Discrete Stochastic Optimization for Public Health Interventions with
Constraints [1.8275108630751844]
This paper addresses aspects of the 2009 H1N1 and the COVID-19 pandemics with the spread of disease modeled by the open source Monte Carlo simulations.
The objective of the optimization is to determine the best combination of intervention strategies so as to result in minimal economic loss to society.
arXiv Detail & Related papers (2022-06-27T21:21:25Z) - Sequential Information Design: Markov Persuasion Process and Its
Efficient Reinforcement Learning [156.5667417159582]
This paper proposes a novel model of sequential information design, namely the Markov persuasion processes (MPPs)
Planning in MPPs faces the unique challenge in finding a signaling policy that is simultaneously persuasive to the myopic receivers and inducing the optimal long-term cumulative utilities of the sender.
We design a provably efficient no-regret learning algorithm, the Optimism-Pessimism Principle for Persuasion Process (OP4), which features a novel combination of both optimism and pessimism principles.
arXiv Detail & Related papers (2022-02-22T05:41:43Z) - Building a Foundation for Data-Driven, Interpretable, and Robust Policy
Design using the AI Economist [67.08543240320756]
We show that the AI Economist framework enables effective, flexible, and interpretable policy design using two-level reinforcement learning and data-driven simulations.
We find that log-linear policies trained using RL significantly improve social welfare, based on both public health and economic outcomes, compared to past outcomes.
arXiv Detail & Related papers (2021-08-06T01:30:41Z) - Machine Learning-Powered Mitigation Policy Optimization in
Epidemiological Models [33.88734751290751]
We propose a new approach for obtaining optimal policy recommendations based on epidemiological models.
We find that such a look-ahead strategy infers non-trivial policies that adhere well to the constraints specified.
arXiv Detail & Related papers (2020-10-16T16:27:17Z) - EpidemiOptim: A Toolbox for the Optimization of Control Policies in
Epidemiological Models [12.748861129923348]
EpidemiOptim is a Python toolbox that facilitates collaborations between researchers in epidemiology and optimization.
It turns epidemiological models and cost functions into optimization problems via a standard interface commonly used by optimization practitioners.
We illustrate the use of EpidemiOptim to find optimal policies for dynamical on-off lock-down control under the optimization of death toll and economic recess.
arXiv Detail & Related papers (2020-10-09T09:25:41Z) - Epidemiologically and Socio-economically Optimal Policies via Bayesian
Optimization [4.910276496877174]
Mass public quarantining, colloquially known as a lock-down, is a non-pharmaceutical intervention to check spread of disease.
This paper presents ESOP, a novel application of active machine learning techniques using Bayesian optimization, that interacts with an epidemiological model to arrive at lock-down schedules that optimally balance public health benefits and socio-economic downsides of reduced economic activity during lock-down periods.
arXiv Detail & Related papers (2020-05-22T16:11:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.