Algorithm as Experiment: Machine Learning, Market Design, and Policy
Eligibility Rules
- URL: http://arxiv.org/abs/2104.12909v6
- Date: Wed, 6 Dec 2023 01:27:41 GMT
- Title: Algorithm as Experiment: Machine Learning, Market Design, and Policy
Eligibility Rules
- Authors: Yusuke Narita and Kohei Yata
- Abstract summary: We develop a treatment-effect estimator using instruments for a class of and deterministic algorithms.
We apply our estimator to evaluate the Coronavirus Aid, Relief, and Economic Security Act.
The funding is shown to have little effect on COVID-19-related hospital activities.
- Score: 10.134708060109404
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Algorithms make a growing portion of policy and business decisions. We
develop a treatment-effect estimator using algorithmic decisions as instruments
for a class of stochastic and deterministic algorithms. Our estimator is
consistent and asymptotically normal for well-defined causal effects. A special
case of our setup is multidimensional regression discontinuity designs with
complex boundaries. We apply our estimator to evaluate the Coronavirus Aid,
Relief, and Economic Security Act, which allocated many billions of dollars
worth of relief funding to hospitals via an algorithmic rule. The funding is
shown to have little effect on COVID-19-related hospital activities. Naive
estimates exhibit selection bias.
Related papers
- Likelihood Ratio Confidence Sets for Sequential Decision Making [51.66638486226482]
We revisit the likelihood-based inference principle and propose to use likelihood ratios to construct valid confidence sequences.
Our method is especially suitable for problems with well-specified likelihoods.
We show how to provably choose the best sequence of estimators and shed light on connections to online convex optimization.
arXiv Detail & Related papers (2023-11-08T00:10:21Z) - Bayesian Safe Policy Learning with Chance Constrained Optimization: Application to Military Security Assessment during the Vietnam War [0.0]
We investigate whether it would have been possible to improve a security assessment algorithm employed during the Vietnam War.
This empirical application raises several methodological challenges that frequently arise in high-stakes algorithmic decision-making.
arXiv Detail & Related papers (2023-07-17T20:59:50Z) - Planning Multiple Epidemic Interventions with Reinforcement Learning [7.51289645756884]
An optimal plan will curb an epidemic with minimal loss of life, disease burden, and economic cost.
Finding an optimal plan is an intractable computational problem in realistic settings.
We apply state-of-the-art actor-critic reinforcement learning algorithms to search for plans that minimize overall costs.
arXiv Detail & Related papers (2023-01-30T11:51:24Z) - Nearly Optimal Latent State Decoding in Block MDPs [74.51224067640717]
In episodic Block MDPs, the decision maker has access to rich observations or contexts generated from a small number of latent states.
We are first interested in estimating the latent state decoding function based on data generated under a fixed behavior policy.
We then study the problem of learning near-optimal policies in the reward-free framework.
arXiv Detail & Related papers (2022-08-17T18:49:53Z) - Safe Online Bid Optimization with Return-On-Investment and Budget
Constraints subject to Uncertainty [87.81197574939355]
We study the nature of both the optimization and learning problems.
We provide an algorithm, namely GCB, guaranteeing sublinear regret at the cost of a potentially linear number of constraints violations.
More interestingly, we provide an algorithm, namely GCB_safe(psi,phi), guaranteeing both sublinear pseudo-regret and safety w.h.p. at the cost of accepting tolerances psi and phi.
arXiv Detail & Related papers (2022-01-18T17:24:20Z) - A Reinforcement Learning Approach to the Stochastic Cutting Stock
Problem [0.0]
We propose a formulation of the cutting stock problem as a discounted infinite-horizon decision process.
An optimal solution corresponds to a policy that associates each state with a decision and minimizes the expected total cost.
arXiv Detail & Related papers (2021-09-20T14:47:54Z) - Bayesian decision-making under misspecified priors with applications to
meta-learning [64.38020203019013]
Thompson sampling and other sequential decision-making algorithms are popular approaches to tackle explore/exploit trade-offs in contextual bandits.
We show that performance degrades gracefully with misspecified priors.
arXiv Detail & Related papers (2021-07-03T23:17:26Z) - Pricing Algorithmic Insurance [3.705785916791345]
We introduce the concept of algorithmic insurance and present a quantitative framework to enable the pricing of the derived insurance contracts.
Our approach outlines how properties of the model, such as accuracy, interpretability and generalizability, can influence the insurance contract evaluation.
arXiv Detail & Related papers (2021-06-01T22:32:02Z) - Stochastic Intervention for Causal Effect Estimation [7.015556609676951]
We propose a new propensity score and intervention effect estimator (SIE) to estimate intervention effect.
We also design a customized genetic algorithm specific to intervention effect (Ge-SIO) with the aim of providing causal evidence for decision making.
Our proposed measures and algorithms can achieve a significant performance lift in comparison with state-of-the-art baselines.
arXiv Detail & Related papers (2021-05-27T01:12:03Z) - Policy Gradient for Continuing Tasks in Non-stationary Markov Decision
Processes [112.38662246621969]
Reinforcement learning considers the problem of finding policies that maximize an expected cumulative reward in a Markov decision process with unknown transition probabilities.
We compute unbiased navigation gradients of the value function which we use as ascent directions to update the policy.
A major drawback of policy gradient-type algorithms is that they are limited to episodic tasks unless stationarity assumptions are imposed.
arXiv Detail & Related papers (2020-10-16T15:15:42Z) - Large-Scale Methods for Distributionally Robust Optimization [53.98643772533416]
We prove that our algorithms require a number of evaluations gradient independent of training set size and number of parameters.
Experiments on MNIST and ImageNet confirm the theoretical scaling of our algorithms, which are 9--36 times more efficient than full-batch methods.
arXiv Detail & Related papers (2020-10-12T17:41:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.