Higher-Order Orthogonal Causal Learning for Treatment Effect
- URL: http://arxiv.org/abs/2103.11869v1
- Date: Mon, 22 Mar 2021 14:04:13 GMT
- Title: Higher-Order Orthogonal Causal Learning for Treatment Effect
- Authors: Yiyan Huang, Cheuk Hang Leung, Xing Yan, Qi Wu
- Abstract summary: We present an algorithm that enables us to obtain the debiased estimator recovered from the score function.
We also undergo comprehensive experiments to test the power of the estimator we construct from the score function using both the simulated datasets and the real datasets.
- Score: 15.652550362252205
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most existing studies on the double/debiased machine learning method
concentrate on the causal parameter estimation recovering from the first-order
orthogonal score function. In this paper, we will construct the
$k^{\mathrm{th}}$-order orthogonal score function for estimating the average
treatment effect (ATE) and present an algorithm that enables us to obtain the
debiased estimator recovered from the score function. Such a higher-order
orthogonal estimator is more robust to the misspecification of the propensity
score than the first-order one does. Besides, it has the merit of being
applicable with many machine learning methodologies such as Lasso, Random
Forests, Neural Nets, etc. We also undergo comprehensive experiments to test
the power of the estimator we construct from the score function using both the
simulated datasets and the real datasets.
Related papers
- Benchmarking Estimators for Natural Experiments: A Novel Dataset and a Doubly Robust Algorithm [12.201705893125775]
We introduce a novel natural experiment dataset obtained from an early childhood literacy nonprofit.
Applying over 20 established estimators to the dataset produces inconsistent results in evaluating the nonprofit's efficacy.
We create a benchmark to evaluate estimator accuracy using synthetic outcomes.
arXiv Detail & Related papers (2024-09-06T15:44:45Z) - Estimating the Hessian Matrix of Ranking Objectives for Stochastic Learning to Rank with Gradient Boosted Trees [63.18324983384337]
We introduce the first learning to rank method for Gradient Boosted Decision Trees (GBDTs)
Our main contribution is a novel estimator for the second-order derivatives, i.e., the Hessian matrix.
We incorporate our estimator into the existing PL-Rank framework, which was originally designed for first-order derivatives only.
arXiv Detail & Related papers (2024-04-18T13:53:32Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Multiply Robust Estimator Circumvents Hyperparameter Tuning of Neural
Network Models in Causal Inference [0.0]
Multiply Robust (MR) estimator allows us to leverage all the first-step models in a single estimator.
We show that MR is the solution to a broad class of estimating equations, and is also consistent if one of the treatment models is $sqrtn$ consistent.
arXiv Detail & Related papers (2023-07-20T02:31:12Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Adaptive LASSO estimation for functional hidden dynamic geostatistical
model [69.10717733870575]
We propose a novel model selection algorithm based on a penalized maximum likelihood estimator (PMLE) for functional hiddenstatistical models (f-HD)
The algorithm is based on iterative optimisation and uses an adaptive least absolute shrinkage and selector operator (GMSOLAS) penalty function, wherein the weights are obtained by the unpenalised f-HD maximum-likelihood estimators.
arXiv Detail & Related papers (2022-08-10T19:17:45Z) - Diversity Enhanced Active Learning with Strictly Proper Scoring Rules [4.81450893955064]
We study acquisition functions for active learning (AL) for text classification.
We convert the Expected Loss Reduction (ELR) method to estimate the increase in (strictly proper) scores like log probability or negative mean square error.
We show that the use of mean square error and log probability with BEMPS yields robust acquisition functions.
arXiv Detail & Related papers (2021-10-27T05:02:11Z) - Learning a Single Neuron with Bias Using Gradient Descent [53.15475693468925]
We study the fundamental problem of learning a single neuron with a bias term.
We show that this is a significantly different and more challenging problem than the bias-less case.
arXiv Detail & Related papers (2021-06-02T12:09:55Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - Doubly Robust Semiparametric Difference-in-Differences Estimators with
High-Dimensional Data [15.27393561231633]
We propose a doubly robust two-stage semiparametric difference-in-difference estimator for estimating heterogeneous treatment effects.
The first stage allows a general set of machine learning methods to be used to estimate the propensity score.
In the second stage, we derive the rates of convergence for both the parametric parameter and the unknown function.
arXiv Detail & Related papers (2020-09-07T15:14:29Z) - Monotonic Cardinality Estimation of Similarity Selection: A Deep
Learning Approach [22.958342743597044]
We investigate the possibilities of utilizing deep learning for cardinality estimation of similarity selection.
We propose a novel and generic method that can be applied to any data type and distance function.
arXiv Detail & Related papers (2020-02-15T20:22:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.