Minimax Instrumental Variable Regression and $L_2$ Convergence
Guarantees without Identification or Closedness
- URL: http://arxiv.org/abs/2302.05404v1
- Date: Fri, 10 Feb 2023 18:08:49 GMT
- Title: Minimax Instrumental Variable Regression and $L_2$ Convergence
Guarantees without Identification or Closedness
- Authors: Andrew Bennett, Nathan Kallus, Xiaojie Mao, Whitney Newey, Vasilis
Syrgkanis, Masatoshi Uehara
- Abstract summary: We study nonparametric estimation of instrumental variable (IV) regressions.
We propose a new penalized minimax estimator that can converge to a fixed IV solution.
We derive a strong $L$ error rate for our estimator under lax conditions.
- Score: 71.42652863687117
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we study nonparametric estimation of instrumental variable
(IV) regressions. Recently, many flexible machine learning methods have been
developed for instrumental variable estimation. However, these methods have at
least one of the following limitations: (1) restricting the IV regression to be
uniquely identified; (2) only obtaining estimation error rates in terms of
pseudometrics (\emph{e.g.,} projected norm) rather than valid metrics
(\emph{e.g.,} $L_2$ norm); or (3) imposing the so-called closedness condition
that requires a certain conditional expectation operator to be sufficiently
smooth. In this paper, we present the first method and analysis that can avoid
all three limitations, while still permitting general function approximation.
Specifically, we propose a new penalized minimax estimator that can converge to
a fixed IV solution even when there are multiple solutions, and we derive a
strong $L_2$ error rate for our estimator under lax conditions. Notably, this
guarantee only needs a widely-used source condition and realizability
assumptions, but not the so-called closedness condition. We argue that the
source condition and the closedness condition are inherently conflicting, so
relaxing the latter significantly improves upon the existing literature that
requires both conditions. Our estimator can achieve this improvement because it
builds on a novel formulation of the IV estimation problem as a constrained
optimization problem.
Related papers
- Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Regularized DeepIV with Model Selection [72.17508967124081]
Regularized DeepIV (RDIV) regression can converge to the least-norm IV solution.
Our method matches the current state-of-the-art convergence rate.
arXiv Detail & Related papers (2024-03-07T05:38:56Z) - Online non-parametric likelihood-ratio estimation by Pearson-divergence
functional minimization [55.98760097296213]
We introduce a new framework for online non-parametric LRE (OLRE) for the setting where pairs of iid observations $(x_t sim p, x'_t sim q)$ are observed over time.
We provide theoretical guarantees for the performance of the OLRE method along with empirical validation in synthetic experiments.
arXiv Detail & Related papers (2023-11-03T13:20:11Z) - A Primal-Dual Approach to Solving Variational Inequalities with General Constraints [54.62996442406718]
Yang et al. (2023) recently showed how to use first-order gradient methods to solve general variational inequalities.
We prove the convergence of this method and show that the gap function of the last iterate of the method decreases at a rate of $O(frac1sqrtK)$ when the operator is $L$-Lipschitz and monotone.
arXiv Detail & Related papers (2022-10-27T17:59:09Z) - On Well-posedness and Minimax Optimal Rates of Nonparametric Q-function
Estimation in Off-policy Evaluation [1.575865518040625]
We study the off-policy evaluation problem in an infinite-horizon Markov decision process with continuous states and actions.
We recast the $Q$-function estimation into a special form of the nonparametric instrumental variables (NPIV) estimation problem.
arXiv Detail & Related papers (2022-01-17T01:09:38Z) - Stochastic regularized majorization-minimization with weakly convex and
multi-convex surrogates [0.0]
We show that the first optimality gap of the proposed algorithm decays at the rate of the expected loss for various methods under nontens dependent data setting.
We obtain first convergence point for various methods under nontens dependent data setting.
arXiv Detail & Related papers (2022-01-05T15:17:35Z) - Off-Policy Interval Estimation with Lipschitz Value Iteration [29.232245317776723]
We propose a provably correct method for obtaining interval bounds for off-policy evaluation in a general continuous setting.
We go on to introduce a Lipschitz value iteration method to monotonically tighten the interval, which is simple yet efficient and provably convergent.
arXiv Detail & Related papers (2020-10-29T07:25:56Z) - Minimax-Optimal Off-Policy Evaluation with Linear Function Approximation [49.502277468627035]
This paper studies the statistical theory of batch data reinforcement learning with function approximation.
Consider the off-policy evaluation problem, which is to estimate the cumulative value of a new target policy from logged history.
arXiv Detail & Related papers (2020-02-21T19:20:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.