Learning Deep Features in Instrumental Variable Regression
- URL: http://arxiv.org/abs/2010.07154v4
- Date: Tue, 27 Jun 2023 10:20:45 GMT
- Title: Learning Deep Features in Instrumental Variable Regression
- Authors: Liyuan Xu, Yutian Chen, Siddarth Srinivasan, Nando de Freitas, Arnaud
Doucet, Arthur Gretton
- Abstract summary: In IV regression, learning proceeds in two stages: stage 1 performs linear regression from the instrument to the treatment; and stage 2 performs linear regression from the treatment to the outcome, conditioned on the instrument.
We propose a novel method, deep feature instrumental variable regression (DFIV), to address the case where relations between instruments, treatments, and outcomes may be nonlinear.
- Score: 42.085253974990046
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Instrumental variable (IV) regression is a standard strategy for learning
causal relationships between confounded treatment and outcome variables from
observational data by utilizing an instrumental variable, which affects the
outcome only through the treatment. In classical IV regression, learning
proceeds in two stages: stage 1 performs linear regression from the instrument
to the treatment; and stage 2 performs linear regression from the treatment to
the outcome, conditioned on the instrument. We propose a novel method, deep
feature instrumental variable regression (DFIV), to address the case where
relations between instruments, treatments, and outcomes may be nonlinear. In
this case, deep neural nets are trained to define informative nonlinear
features on the instruments and treatments. We propose an alternating training
regime for these features to ensure good end-to-end performance when composing
stages 1 and 2, thus obtaining highly flexible feature maps in a
computationally efficient manner. DFIV outperforms recent state-of-the-art
methods on challenging IV benchmarks, including settings involving high
dimensional image data. DFIV also exhibits competitive performance in
off-policy policy evaluation for reinforcement learning, which can be
understood as an IV regression task.
Related papers
- Geometry-Aware Instrumental Variable Regression [56.16884466478886]
We propose a transport-based IV estimator that takes into account the geometry of the data manifold through data-derivative information.
We provide a simple plug-and-play implementation of our method that performs on par with related estimators in standard settings.
arXiv Detail & Related papers (2024-05-19T17:49:33Z) - Learning Decision Policies with Instrumental Variables through Double Machine Learning [16.842233444365764]
A common issue in learning decision-making policies in data-rich settings is spurious correlations in the offline dataset.
We propose DML-IV, a non-linear IV regression method that reduces the bias in two-stage IV regressions.
It outperforms state-of-the-art IV regression methods on IV regression benchmarks and learns high-performing policies in the presence of instruments.
arXiv Detail & Related papers (2024-05-14T10:55:04Z) - Regularized DeepIV with Model Selection [72.17508967124081]
Regularized DeepIV (RDIV) regression can converge to the least-norm IV solution.
Our method matches the current state-of-the-art convergence rate.
arXiv Detail & Related papers (2024-03-07T05:38:56Z) - Statistically Efficient Variance Reduction with Double Policy Estimation
for Off-Policy Evaluation in Sequence-Modeled Reinforcement Learning [53.97273491846883]
We propose DPE: an RL algorithm that blends offline sequence modeling and offline reinforcement learning with Double Policy Estimation.
We validate our method in multiple tasks of OpenAI Gym with D4RL benchmarks.
arXiv Detail & Related papers (2023-08-28T20:46:07Z) - Near-optimal Offline Reinforcement Learning with Linear Representation:
Leveraging Variance Information with Pessimism [65.46524775457928]
offline reinforcement learning seeks to utilize offline/historical data to optimize sequential decision-making strategies.
We study the statistical limits of offline reinforcement learning with linear model representations.
arXiv Detail & Related papers (2022-03-11T09:00:12Z) - Jump Interval-Learning for Individualized Decision Making [21.891586204541877]
We propose a jump interval-learning to develop an individualized interval-valued decision rule (I2DR)
Unlike IDRs that recommend a single treatment, the proposed I2DR yields an interval of treatment options for each individual.
arXiv Detail & Related papers (2021-11-17T03:29:59Z) - Improving Inference from Simple Instruments through Compliance
Estimation [0.0]
Instrumental variables (IV) regression is widely used to estimate causal treatment effects in settings where receipt of treatment is not fully random.
While IV can recover consistent treatment effect estimates, they are often noisy.
We study how to improve the efficiency of IV estimates by exploiting the predictable variation in the strength of the instrument.
arXiv Detail & Related papers (2021-08-08T20:18:34Z) - Instrumental Variable Value Iteration for Causal Offline Reinforcement Learning [107.70165026669308]
In offline reinforcement learning (RL) an optimal policy is learned solely from a priori collected observational data.
We study a confounded Markov decision process where the transition dynamics admit an additive nonlinear functional form.
We propose a provably efficient IV-aided Value Iteration (IVVI) algorithm based on a primal-dual reformulation of the conditional moment restriction.
arXiv Detail & Related papers (2021-02-19T13:01:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.