State Relevance for Off-Policy Evaluation
- URL: http://arxiv.org/abs/2109.06310v1
- Date: Mon, 13 Sep 2021 20:40:55 GMT
- Title: State Relevance for Off-Policy Evaluation
- Authors: Simon P. Shen, Yecheng Jason Ma, Omer Gottesman, Finale Doshi-Velez
- Abstract summary: We introduce Omitting-States-Irrelevant-to-Return Importance Sampling (OSIRIS), an estimator which reduces variance by strategically omitting likelihood ratios associated with certain states.
We formalize the conditions under which OSIRIS is unbiased and has lower variance than ordinary importance sampling.
- Score: 29.891687579606277
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Importance sampling-based estimators for off-policy evaluation (OPE) are
valued for their simplicity, unbiasedness, and reliance on relatively few
assumptions. However, the variance of these estimators is often high,
especially when trajectories are of different lengths. In this work, we
introduce Omitting-States-Irrelevant-to-Return Importance Sampling (OSIRIS), an
estimator which reduces variance by strategically omitting likelihood ratios
associated with certain states. We formalize the conditions under which OSIRIS
is unbiased and has lower variance than ordinary importance sampling, and we
demonstrate these properties empirically.
Related papers
- Marginal Density Ratio for Off-Policy Evaluation in Contextual Bandits [41.91108406329159]
Off-Policy Evaluation (OPE) in contextual bandits is crucial for assessing new policies using existing data without costly experimentation.
We introduce a new OPE estimator for contextual bandits, the Marginal Ratio (MR) estimator, which focuses on the shift in the marginal distribution of outcomes $Y$ instead of the policies themselves.
arXiv Detail & Related papers (2023-12-03T17:04:57Z) - A Tale of Sampling and Estimation in Discounted Reinforcement Learning [50.43256303670011]
We present a minimax lower bound on the discounted mean estimation problem.
We show that estimating the mean by directly sampling from the discounted kernel of the Markov process brings compelling statistical properties.
arXiv Detail & Related papers (2023-04-11T09:13:17Z) - Uncertainty-Aware Instance Reweighting for Off-Policy Learning [63.31923483172859]
We propose a Uncertainty-aware Inverse Propensity Score estimator (UIPS) for improved off-policy learning.
Experiment results on synthetic and three real-world recommendation datasets demonstrate the advantageous sample efficiency of the proposed UIPS estimator.
arXiv Detail & Related papers (2023-03-11T11:42:26Z) - Improved Policy Evaluation for Randomized Trials of Algorithmic Resource
Allocation [54.72195809248172]
We present a new estimator leveraging our proposed novel concept, that involves retrospective reshuffling of participants across experimental arms at the end of an RCT.
We prove theoretically that such an estimator is more accurate than common estimators based on sample means.
arXiv Detail & Related papers (2023-02-06T05:17:22Z) - Low Variance Off-policy Evaluation with State-based Importance Sampling [21.727827944373793]
This paper proposes state-based importance sampling estimators which reduce the variance by dropping certain states from the computation of the importance weight.
Experiments in four domains show that state-based methods consistently yield reduced variance and improved accuracy compared to their traditional counterparts.
arXiv Detail & Related papers (2022-12-07T19:56:11Z) - Off-policy evaluation for learning-to-rank via interpolating the
item-position model and the position-based model [83.83064559894989]
A critical need for industrial recommender systems is the ability to evaluate recommendation policies offline, before deploying them to production.
We develop a new estimator that mitigates the problems of the two most popular off-policy estimators for rankings.
In particular, the new estimator, called INTERPOL, addresses the bias of a potentially misspecified position-based model.
arXiv Detail & Related papers (2022-10-15T17:22:30Z) - The role of the geometric mean in case-control studies [4.38301148531795]
We describe how to partially identify, estimate, and do inference on the geometric odds ratio under outcome-dependent sampling.
Our proposed estimator is based on the efficient influence function and therefore has doubly robust-style properties.
arXiv Detail & Related papers (2022-07-19T01:42:52Z) - Off-Policy Evaluation for Large Action Spaces via Embeddings [36.42838320396534]
Off-policy evaluation (OPE) in contextual bandits has seen rapid adoption in real-world systems.
Existing OPE estimators degrade severely when the number of actions is large.
We propose a new OPE estimator that leverages marginalized importance weights when action embeddings provide structure in the action space.
arXiv Detail & Related papers (2022-02-13T14:00:09Z) - SOPE: Spectrum of Off-Policy Estimators [40.15700429288981]
We show the existence of a spectrum of estimators whose endpoints are SIS and IS.
We provide empirical evidence that estimators in this spectrum can be used to trade-off between the bias and variance of IS and SIS.
arXiv Detail & Related papers (2021-11-06T18:29:21Z) - Universal Off-Policy Evaluation [64.02853483874334]
We take the first steps towards a universal off-policy estimator (UnO)
We use UnO for estimating and simultaneously bounding the mean, variance, quantiles/median, inter-quantile range, CVaR, and the entire cumulative distribution of returns.
arXiv Detail & Related papers (2021-04-26T18:54:31Z) - Deconfounding Scores: Feature Representations for Causal Effect
Estimation with Weak Overlap [140.98628848491146]
We introduce deconfounding scores, which induce better overlap without biasing the target of estimation.
We show that deconfounding scores satisfy a zero-covariance condition that is identifiable in observed data.
In particular, we show that this technique could be an attractive alternative to standard regularizations.
arXiv Detail & Related papers (2021-04-12T18:50:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.