One-Step Estimation of Differentiable Hilbert-Valued Parameters
- URL: http://arxiv.org/abs/2303.16711v3
- Date: Wed, 27 Sep 2023 00:07:28 GMT
- Title: One-Step Estimation of Differentiable Hilbert-Valued Parameters
- Authors: Alex Luedtke and Incheoul Chung
- Abstract summary: We present estimators for smooth Hilbert-valued parameters, where smoothness is characterized by a pathwise differentiability condition.
These estimators correspond to generalizations of cross-fitted one-step estimators based on Hilbert-valued efficient influence functions.
- Score: 2.0305676256390934
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present estimators for smooth Hilbert-valued parameters, where smoothness
is characterized by a pathwise differentiability condition. When the parameter
space is a reproducing kernel Hilbert space, we provide a means to obtain
efficient, root-n rate estimators and corresponding confidence sets. These
estimators correspond to generalizations of cross-fitted one-step estimators
based on Hilbert-valued efficient influence functions. We give theoretical
guarantees even when arbitrary estimators of nuisance functions are used,
including those based on machine learning techniques. We show that these
results naturally extend to Hilbert spaces that lack a reproducing kernel, as
long as the parameter has an efficient influence function. However, we also
uncover the unfortunate fact that, when there is no reproducing kernel, many
interesting parameters fail to have an efficient influence function, even
though they are pathwise differentiable. To handle these cases, we propose a
regularized one-step estimator and associated confidence sets. We also show
that pathwise differentiability, which is a central requirement of our
approach, holds in many cases. Specifically, we provide multiple examples of
pathwise differentiable parameters and develop corresponding estimators and
confidence sets. Among these examples, four are particularly relevant to
ongoing research by the causal inference community: the counterfactual density
function, dose-response function, conditional average treatment effect
function, and counterfactual kernel mean embedding.
Related papers
- Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
Temporal Difference (TD) learning, arguably the most widely used for policy evaluation, serves as a natural framework for this purpose.
In this paper, we study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation, and obtain three significant improvements over existing results.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - Statistical Inference of Optimal Allocations I: Regularities and their Implications [3.904240476752459]
We first derive Hadamard differentiability of the value function through a detailed analysis of the general properties of the sorting operator.
Building on our Hadamard differentiability results, we demonstrate how the functional delta method can be used to directly derive the properties of the value function process.
arXiv Detail & Related papers (2024-03-27T04:39:13Z) - Kernel-based off-policy estimation without overlap: Instance optimality
beyond semiparametric efficiency [53.90687548731265]
We study optimal procedures for estimating a linear functional based on observational data.
For any convex and symmetric function class $mathcalF$, we derive a non-asymptotic local minimax bound on the mean-squared error.
arXiv Detail & Related papers (2023-01-16T02:57:37Z) - Data-Driven Influence Functions for Optimization-Based Causal Inference [105.5385525290466]
We study a constructive algorithm that approximates Gateaux derivatives for statistical functionals by finite differencing.
We study the case where probability distributions are not known a priori but need to be estimated from data.
arXiv Detail & Related papers (2022-08-29T16:16:22Z) - Inference on Strongly Identified Functionals of Weakly Identified
Functions [71.42652863687117]
We study a novel condition for the functional to be strongly identified even when the nuisance function is not.
We propose penalized minimax estimators for both the primary and debiasing nuisance functions.
arXiv Detail & Related papers (2022-08-17T13:38:31Z) - Experimental Design for Linear Functionals in Reproducing Kernel Hilbert
Spaces [102.08678737900541]
We provide algorithms for constructing bias-aware designs for linear functionals.
We derive non-asymptotic confidence sets for fixed and adaptive designs under sub-Gaussian noise.
arXiv Detail & Related papers (2022-05-26T20:56:25Z) - Minimax Kernel Machine Learning for a Class of Doubly Robust Functionals [16.768606469968113]
We consider a class of doubly robust moment functions originally introduced in (Robins et al., 2008)
We demonstrate that this moment function can be used to construct estimating equations for the nuisance functions.
The convergence rates of the nuisance functions are analyzed using the modern techniques in statistical learning theory.
arXiv Detail & Related papers (2021-04-07T05:52:15Z) - Causal Inference Under Unmeasured Confounding With Negative Controls: A
Minimax Learning Approach [84.29777236590674]
We study the estimation of causal parameters when not all confounders are observed and instead negative controls are available.
Recent work has shown how these can enable identification and efficient estimation via two so-called bridge functions.
arXiv Detail & Related papers (2021-03-25T17:59:19Z) - Adversarial Estimation of Riesz Representers [21.510036777607397]
We propose an adversarial framework to estimate the Riesz representer using general function spaces.
We prove a nonasymptotic mean square rate in terms of an abstract quantity called the critical radius, then specialize it for neural networks, random forests, and reproducing kernel Hilbert spaces as leading cases.
arXiv Detail & Related papers (2020-12-30T19:46:57Z) - Equivalence of Convergence Rates of Posterior Distributions and Bayes
Estimators for Functions and Nonparametric Functionals [4.375582647111708]
We study the posterior contraction rates of a Bayesian method with Gaussian process priors in nonparametric regression.
For a general class of kernels, we establish convergence rates of the posterior measure of the regression function and its derivatives.
Our proof shows that, under certain conditions, to any convergence rate of Bayes estimators there corresponds the same convergence rate of the posterior distributions.
arXiv Detail & Related papers (2020-11-27T19:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.