Counterfactual Learning with Multioutput Deep Kernels
- URL: http://arxiv.org/abs/2211.11119v1
- Date: Sun, 20 Nov 2022 23:28:41 GMT
- Title: Counterfactual Learning with Multioutput Deep Kernels
- Authors: Alberto Caron, Gianluca Baio, Ioanna Manolopoulou
- Abstract summary: In this paper, we address the challenge of performing counterfactual inference with observational data.
We present a general class of counterfactual multi-task deep kernels models that estimate causal effects and learn policies proficiently.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the challenge of performing counterfactual
inference with observational data via Bayesian nonparametric regression
adjustment, with a focus on high-dimensional settings featuring multiple
actions and multiple correlated outcomes. We present a general class of
counterfactual multi-task deep kernels models that estimate causal effects and
learn policies proficiently thanks to their sample efficiency gains, while
scaling well with high dimensions. In the first part of the work, we rely on
Structural Causal Models (SCM) to formally introduce the setup and the problem
of identifying counterfactual quantities under observed confounding. We then
discuss the benefits of tackling the task of causal effects estimation via
stacked coregionalized Gaussian Processes and Deep Kernels. Finally, we
demonstrate the use of the proposed methods on simulated experiments that span
individual causal effects estimation, off-policy evaluation and optimization.
Related papers
- Estimation and Inference for Causal Functions with Multiway Clustered Data [6.988496457312806]
This paper proposes methods of estimation and uniform inference for a general class of causal functions.
The causal function is identified as a conditional expectation of an adjusted (Neyman-orthogonal) signal.
We apply the proposed methods to analyze the causal relationship between levels in Africa and the historical slave trade.
arXiv Detail & Related papers (2024-09-10T17:17:53Z) - Denoising Pre-Training and Customized Prompt Learning for Efficient Multi-Behavior Sequential Recommendation [69.60321475454843]
We propose DPCPL, the first pre-training and prompt-tuning paradigm tailored for Multi-Behavior Sequential Recommendation.
In the pre-training stage, we propose a novel Efficient Behavior Miner (EBM) to filter out the noise at multiple time scales.
Subsequently, we propose to tune the pre-trained model in a highly efficient manner with the proposed Customized Prompt Learning (CPL) module.
arXiv Detail & Related papers (2024-08-21T06:48:38Z) - Bayesian Off-Policy Evaluation and Learning for Large Action Spaces [14.203316003782604]
In interactive systems, actions are often correlated, presenting an opportunity for more sample-efficient off-policy evaluation and learning.
We introduce a unified Bayesian framework to capture these correlations through structured and informative priors.
We propose sDM, a generic Bayesian approach for OPE and OPL, grounded in both algorithmic and theoretical foundations.
arXiv Detail & Related papers (2024-02-22T16:09:45Z) - ACE : Off-Policy Actor-Critic with Causality-Aware Entropy Regularization [52.5587113539404]
We introduce a causality-aware entropy term that effectively identifies and prioritizes actions with high potential impacts for efficient exploration.
Our proposed algorithm, ACE: Off-policy Actor-critic with Causality-aware Entropy regularization, demonstrates a substantial performance advantage across 29 diverse continuous control tasks.
arXiv Detail & Related papers (2024-02-22T13:22:06Z) - Benchmarking Bayesian Causal Discovery Methods for Downstream Treatment
Effect Estimation [137.3520153445413]
A notable gap exists in the evaluation of causal discovery methods, where insufficient emphasis is placed on downstream inference.
We evaluate seven established baseline causal discovery methods including a newly proposed method based on GFlowNets.
The results of our study demonstrate that some of the algorithms studied are able to effectively capture a wide range of useful and diverse ATE modes.
arXiv Detail & Related papers (2023-07-11T02:58:10Z) - Estimating Causal Effects using a Multi-task Deep Ensemble [4.268861137988059]
Causal Multi-task Deep Ensemble (CMDE) is a novel framework that learns both shared and group-specific information from the study population.
We evaluate our method across various types of datasets and tasks and find that CMDE outperforms state-of-the-art methods on a majority of these tasks.
arXiv Detail & Related papers (2023-01-26T19:05:18Z) - Counterfactual Maximum Likelihood Estimation for Training Deep Networks [83.44219640437657]
Deep learning models are prone to learning spurious correlations that should not be learned as predictive clues.
We propose a causality-based training framework to reduce the spurious correlations caused by observable confounders.
We conduct experiments on two real-world tasks: Natural Language Inference (NLI) and Image Captioning.
arXiv Detail & Related papers (2021-06-07T17:47:16Z) - Multi-Stage Influence Function [97.19210942277354]
We develop a multi-stage influence function score to track predictions from a finetuned model all the way back to the pretraining data.
We study two different scenarios with the pretrained embeddings fixed or updated in the finetuning tasks.
arXiv Detail & Related papers (2020-07-17T16:03:11Z) - Influence Functions in Deep Learning Are Fragile [52.31375893260445]
influence functions approximate the effect of samples in test-time predictions.
influence estimates are fairly accurate for shallow networks.
Hessian regularization is important to get highquality influence estimates.
arXiv Detail & Related papers (2020-06-25T18:25:59Z) - Consistency of Empirical Bayes And Kernel Flow For Hierarchical
Parameter Estimation [18.852474191260445]
This paper studies two paradigms of learning hierarchical parameters.
One is from the probabilistic Bayesian perspective, in particular, the empirical Bayes approach.
The other is from the deterministic and approximation theoretic view, and in particular the kernel flow algorithm.
arXiv Detail & Related papers (2020-05-22T20:17:48Z) - Nonparametric inference for interventional effects with multiple
mediators [0.0]
We provide theory that allows for more flexible, possibly machine learning-based, estimation techniques.
We demonstrate multiple robustness properties of the proposed estimators.
Our work thus provides a means of leveraging modern statistical learning techniques in estimation of interventional mediation effects.
arXiv Detail & Related papers (2020-01-16T19:05:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.